Commit 96381206 by Venkatesh Seetharam

Fix rebase issues. Contributed by Venkatesh Seetharam

parent a92298c5
...@@ -45,12 +45,6 @@ ...@@ -45,12 +45,6 @@
<version>2.2.2</version> <version>2.2.2</version>
</dependency> </dependency>
<dependency> <dependency>
<groupId>junit</groupId>
<artifactId>junit</artifactId>
<version>4.10</version>
<scope>test</scope>
</dependency>
<dependency>
<groupId>org.apache.hadoop.metadata</groupId> <groupId>org.apache.hadoop.metadata</groupId>
<artifactId>metadata-typesystem</artifactId> <artifactId>metadata-typesystem</artifactId>
</dependency> </dependency>
...@@ -75,23 +69,8 @@ ...@@ -75,23 +69,8 @@
<artifactId>log4j</artifactId> <artifactId>log4j</artifactId>
</dependency> </dependency>
<dependency> <dependency>
<groupId>com.sun.jersey</groupId> <groupId>org.testng</groupId>
<artifactId>jersey-core</artifactId> <artifactId>testng</artifactId>
</dependency>
<dependency>
<groupId>com.sun.jersey</groupId>
<artifactId>jersey-server</artifactId>
</dependency>
<dependency>
<groupId>com.sun.jersey</groupId>
<artifactId>jersey-client</artifactId>
</dependency>
<dependency>
<groupId>com.sun.jersey</groupId>
<artifactId>jersey-json</artifactId>
</dependency> </dependency>
</dependencies> </dependencies>
......
...@@ -32,18 +32,17 @@ import org.apache.hadoop.metadata.repository.graph.GraphHelper; ...@@ -32,18 +32,17 @@ import org.apache.hadoop.metadata.repository.graph.GraphHelper;
import org.apache.hadoop.metadata.repository.graph.GraphService; import org.apache.hadoop.metadata.repository.graph.GraphService;
import org.apache.hadoop.metadata.repository.graph.TitanGraphProvider; import org.apache.hadoop.metadata.repository.graph.TitanGraphProvider;
import org.apache.hadoop.metadata.repository.graph.TitanGraphService; import org.apache.hadoop.metadata.repository.graph.TitanGraphService;
import org.junit.After;
import org.junit.Before;
import org.junit.Ignore;
import org.junit.Test;
import org.slf4j.Logger; import org.slf4j.Logger;
import org.slf4j.LoggerFactory; import org.slf4j.LoggerFactory;
import org.testng.annotations.AfterClass;
import org.testng.annotations.BeforeClass;
import org.testng.annotations.Test;
import java.io.BufferedWriter; import java.io.BufferedWriter;
import java.io.FileWriter; import java.io.FileWriter;
import java.util.List; import java.util.List;
@Ignore @Test (enabled = false)
public class HiveGraphRepositoryTest { public class HiveGraphRepositoryTest {
protected HiveTypeSystem hts; protected HiveTypeSystem hts;
...@@ -53,7 +52,7 @@ public class HiveGraphRepositoryTest { ...@@ -53,7 +52,7 @@ public class HiveGraphRepositoryTest {
private static final Logger LOG = private static final Logger LOG =
LoggerFactory.getLogger(HiveGraphRepositoryTest.class); LoggerFactory.getLogger(HiveGraphRepositoryTest.class);
@Before @BeforeClass
public void setup() throws ConfigurationException, MetadataException { public void setup() throws ConfigurationException, MetadataException {
gs = new TitanGraphService(new TitanGraphProvider()); gs = new TitanGraphService(new TitanGraphProvider());
...@@ -61,7 +60,7 @@ public class HiveGraphRepositoryTest { ...@@ -61,7 +60,7 @@ public class HiveGraphRepositoryTest {
hts = HiveTypeSystem.getInstance(); hts = HiveTypeSystem.getInstance();
} }
@After @AfterClass
public void tearDown() { public void tearDown() {
Graph graph = gs.getBlueprintsGraph(); Graph graph = gs.getBlueprintsGraph();
System.out.println("*******************Graph Dump****************************"); System.out.println("*******************Graph Dump****************************");
...@@ -77,7 +76,7 @@ public class HiveGraphRepositoryTest { ...@@ -77,7 +76,7 @@ public class HiveGraphRepositoryTest {
System.out.println("*******************Graph Dump****************************"); System.out.println("*******************Graph Dump****************************");
} }
@Test @Test (enabled = false)
public void testHiveImport() throws Exception { public void testHiveImport() throws Exception {
HiveImporter hImporter = new HiveImporter(repository, hts, new HiveMetaStoreClient(new HiveConf())); HiveImporter hImporter = new HiveImporter(repository, hts, new HiveMetaStoreClient(new HiveConf()));
......
...@@ -18,8 +18,6 @@ ...@@ -18,8 +18,6 @@
package org.apache.hadoop.metadata.hivetypes; package org.apache.hadoop.metadata.hivetypes;
import org.apache.commons.logging.Log;
import org.apache.commons.logging.LogFactory;
import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.conf.HiveConf;
import org.apache.hadoop.hive.metastore.HiveMetaStoreClient; import org.apache.hadoop.hive.metastore.HiveMetaStoreClient;
import org.apache.hadoop.hive.metastore.api.MetaException; import org.apache.hadoop.hive.metastore.api.MetaException;
...@@ -28,18 +26,16 @@ import org.apache.hadoop.metadata.MetadataException; ...@@ -28,18 +26,16 @@ import org.apache.hadoop.metadata.MetadataException;
import org.apache.hadoop.metadata.storage.Id; import org.apache.hadoop.metadata.storage.Id;
import org.apache.hadoop.metadata.storage.memory.MemRepository; import org.apache.hadoop.metadata.storage.memory.MemRepository;
import org.apache.hadoop.metadata.types.TypeSystem; import org.apache.hadoop.metadata.types.TypeSystem;
import org.junit.Before;
import org.junit.Ignore;
import org.junit.Test;
import org.slf4j.Logger; import org.slf4j.Logger;
import org.slf4j.LoggerFactory; import org.slf4j.LoggerFactory;
import org.testng.annotations.BeforeClass;
import org.testng.annotations.Test;
import java.io.BufferedWriter; import java.io.BufferedWriter;
import java.io.FileOutputStream;
import java.io.FileWriter; import java.io.FileWriter;
import java.io.IOException; import java.io.IOException;
@Ignore @Test (enabled = false)
public class HiveTypeSystemTest { public class HiveTypeSystemTest {
protected MemRepository mr; protected MemRepository mr;
...@@ -48,7 +44,7 @@ public class HiveTypeSystemTest { ...@@ -48,7 +44,7 @@ public class HiveTypeSystemTest {
private static final Logger LOG = private static final Logger LOG =
LoggerFactory.getLogger(HiveTypeSystemTest.class); LoggerFactory.getLogger(HiveTypeSystemTest.class);
@Before @BeforeClass
public void setup() throws MetadataException { public void setup() throws MetadataException {
TypeSystem ts = TypeSystem.getInstance(); TypeSystem ts = TypeSystem.getInstance();
...@@ -57,7 +53,7 @@ public class HiveTypeSystemTest { ...@@ -57,7 +53,7 @@ public class HiveTypeSystemTest {
hts = HiveTypeSystem.getInstance(); hts = HiveTypeSystem.getInstance();
} }
@Test @Test (enabled = false)
public void testHiveImport() throws MetaException, MetadataException, IOException { public void testHiveImport() throws MetaException, MetadataException, IOException {
HiveImporter hImporter = new HiveImporter(mr, hts, new HiveMetaStoreClient(new HiveConf())); HiveImporter hImporter = new HiveImporter(mr, hts, new HiveMetaStoreClient(new HiveConf()));
...@@ -87,5 +83,4 @@ public class HiveTypeSystemTest { ...@@ -87,5 +83,4 @@ public class HiveTypeSystemTest {
bw.flush(); bw.flush();
bw.close(); bw.close();
} }
} }
\ No newline at end of file
package org.apache.hadoop.metadata.web.resources; package org.apache.hadoop.metadata.web.resources;
import java.io.IOException; import com.google.gson.Gson;
import java.io.InputStreamReader; import com.google.gson.JsonArray;
import java.io.Reader; import com.google.gson.JsonElement;
import java.io.Writer; import com.google.gson.JsonObject;
import java.util.ArrayList; import com.google.gson.JsonPrimitive;
import java.util.List; import org.apache.hadoop.metadata.MetadataException;
import org.apache.hadoop.metadata.bridge.hivelineage.HiveLineageBridge;
import org.apache.hadoop.metadata.bridge.hivelineage.hook.HiveLineage;
import org.apache.hadoop.metadata.storage.RepositoryException;
import javax.inject.Inject; import javax.inject.Inject;
import javax.inject.Singleton; import javax.inject.Singleton;
...@@ -18,16 +21,9 @@ import javax.ws.rs.PathParam; ...@@ -18,16 +21,9 @@ import javax.ws.rs.PathParam;
import javax.ws.rs.Produces; import javax.ws.rs.Produces;
import javax.ws.rs.core.Context; import javax.ws.rs.core.Context;
import javax.ws.rs.core.MediaType; import javax.ws.rs.core.MediaType;
import java.io.IOException;
import org.apache.hadoop.metadata.bridge.hivelineage.HiveLineageBridge; import java.io.InputStreamReader;
import org.apache.hadoop.metadata.bridge.hivelineage.hook.HiveLineage; import java.io.Reader;
import org.apache.hadoop.metadata.storage.RepositoryException;
import com.google.gson.Gson;
import com.google.gson.JsonArray;
import com.google.gson.JsonElement;
import com.google.gson.JsonObject;
import com.google.gson.JsonPrimitive;
@Path("bridge/hive") @Path("bridge/hive")
@Singleton @Singleton
...@@ -73,10 +69,10 @@ public class HiveLineageResource { ...@@ -73,10 +69,10 @@ public class HiveLineageResource {
@POST @POST
@Consumes(MediaType.APPLICATION_JSON) @Consumes(MediaType.APPLICATION_JSON)
@Produces(MediaType.APPLICATION_JSON) @Produces(MediaType.APPLICATION_JSON)
public JsonElement addLineage(@Context HttpServletRequest request) throws IOException, RepositoryException { public JsonElement addLineage(@Context HttpServletRequest request)
throws IOException, MetadataException {
// create a reader // create a reader
Reader reader = new InputStreamReader(request.getInputStream()); try (Reader reader = new InputStreamReader(request.getInputStream())) {
try {
// deserialize // deserialize
HiveLineage bean = new Gson().fromJson(reader, HiveLineage.class); HiveLineage bean = new Gson().fromJson(reader, HiveLineage.class);
String id = bridge.create(bean); String id = bridge.create(bean);
...@@ -84,9 +80,6 @@ public class HiveLineageResource { ...@@ -84,9 +80,6 @@ public class HiveLineageResource {
JsonObject jo = new JsonObject(); JsonObject jo = new JsonObject();
jo.addProperty("id", id); jo.addProperty("id", id);
return jo; return jo;
} finally {
// be a good citizen
reader.close();
} }
} }
} }
/**
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.hadoop.metadata.repository.graph; package org.apache.hadoop.metadata.repository.graph;
import com.google.common.collect.ImmutableList; import com.google.common.collect.ImmutableList;
...@@ -6,7 +24,7 @@ import com.tinkerpop.blueprints.Edge; ...@@ -6,7 +24,7 @@ import com.tinkerpop.blueprints.Edge;
import com.tinkerpop.blueprints.Vertex; import com.tinkerpop.blueprints.Vertex;
import org.apache.hadoop.metadata.ITypedReferenceableInstance; import org.apache.hadoop.metadata.ITypedReferenceableInstance;
import org.apache.hadoop.metadata.Referenceable; import org.apache.hadoop.metadata.Referenceable;
import org.apache.hadoop.metadata.RepositoryModuleBaseTest; import org.apache.hadoop.metadata.RepositoryMetadataModule;
import org.apache.hadoop.metadata.Struct; import org.apache.hadoop.metadata.Struct;
import org.apache.hadoop.metadata.types.AttributeDefinition; import org.apache.hadoop.metadata.types.AttributeDefinition;
import org.apache.hadoop.metadata.types.ClassType; import org.apache.hadoop.metadata.types.ClassType;
...@@ -20,25 +38,32 @@ import org.apache.hadoop.metadata.types.StructTypeDefinition; ...@@ -20,25 +38,32 @@ import org.apache.hadoop.metadata.types.StructTypeDefinition;
import org.apache.hadoop.metadata.types.TraitType; import org.apache.hadoop.metadata.types.TraitType;
import org.apache.hadoop.metadata.types.TypeSystem; import org.apache.hadoop.metadata.types.TypeSystem;
import org.testng.annotations.BeforeClass; import org.testng.annotations.BeforeClass;
import org.testng.annotations.Guice;
import org.testng.annotations.Test; import org.testng.annotations.Test;
public class GraphRepoMapperTest extends RepositoryModuleBaseTest { import javax.inject.Inject;
@Test
@Guice(modules = RepositoryMetadataModule.class)
public class GraphRepoMapperTest {
private static final String DATABASE_TYPE = "hive_database"; private static final String DATABASE_TYPE = "hive_database";
private static final String DATABASE_NAME = "foo"; private static final String DATABASE_NAME = "foo";
private static final String TABLE_TYPE = "hive_table"; private static final String TABLE_TYPE = "hive_table";
private static final String TABLE_NAME = "bar"; private static final String TABLE_NAME = "bar";
private TitanGraphService titanGraphService; @Inject
private GraphBackedMetadataRepository repositoryService; TitanGraphService titanGraphService;
@Inject
GraphBackedMetadataRepository repositoryService;
private TypeSystem typeSystem; private TypeSystem typeSystem;
@BeforeClass @BeforeClass
public void setUp() throws Exception { public void setUp() throws Exception {
titanGraphService = super.injector.getInstance(TitanGraphService.class); // start the injected graph service
titanGraphService.start(); titanGraphService.start();
// start the injected repository service
repositoryService = super.injector.getInstance(GraphBackedMetadataRepository.class);
repositoryService.start(); repositoryService.start();
typeSystem = TypeSystem.getInstance(); typeSystem = TypeSystem.getInstance();
......
...@@ -60,9 +60,9 @@ public class GraphBackedMetadataRepositoryTest { ...@@ -60,9 +60,9 @@ public class GraphBackedMetadataRepositoryTest {
private static final String ENTITY_TYPE = "Department"; private static final String ENTITY_TYPE = "Department";
@Inject @Inject
TitanGraphService titanGraphService; private TitanGraphService titanGraphService;
@Inject @Inject
GraphBackedMetadataRepository repositoryService; private GraphBackedMetadataRepository repositoryService;
private TypeSystem ts; private TypeSystem ts;
private String guid; private String guid;
......
Markdown is supported
0% or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment