Skip to content
Projects
Groups
Snippets
Help
This project
Loading...
Sign in / Register
Toggle navigation
A
atlas
Project
Overview
Details
Activity
Cycle Analytics
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Charts
Issues
0
Issues
0
List
Board
Labels
Milestones
Merge Requests
0
Merge Requests
0
CI / CD
CI / CD
Pipelines
Jobs
Schedules
Charts
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Charts
Create a new issue
Jobs
Commits
Issue Boards
Open sidebar
dataplatform
atlas
Commits
80674f0a
Commit
80674f0a
authored
May 29, 2015
by
Suma Shivaprasad
Browse files
Options
Browse Files
Download
Plain Diff
Merged with master
parents
018bbd62
25a68075
Show whitespace changes
Inline
Side-by-side
Showing
23 changed files
with
315 additions
and
168 deletions
+315
-168
pom.xml
addons/hive-bridge/pom.xml
+1
-9
HiveHookIT.java
...java/org/apache/hadoop/metadata/hive/hook/HiveHookIT.java
+4
-0
pom.xml
client/pom.xml
+11
-15
pom.xml
pom.xml
+75
-2
GraphTransaction.java
...ain/java/org/apache/hadoop/metadata/GraphTransaction.java
+26
-0
GraphTransactionInterceptor.java
...g/apache/hadoop/metadata/GraphTransactionInterceptor.java
+53
-0
RepositoryMetadataModule.java
.../org/apache/hadoop/metadata/RepositoryMetadataModule.java
+13
-31
HiveLineageService.java
.../apache/hadoop/metadata/discovery/HiveLineageService.java
+6
-0
DefaultGraphPersistenceStrategy.java
...data/discovery/graph/DefaultGraphPersistenceStrategy.java
+8
-23
GraphBackedDiscoveryService.java
...metadata/discovery/graph/GraphBackedDiscoveryService.java
+4
-0
GraphBackedMetadataRepository.java
...adata/repository/graph/GraphBackedMetadataRepository.java
+8
-20
TitanGraphProvider.java
.../hadoop/metadata/repository/graph/TitanGraphProvider.java
+2
-0
GraphBackedTypeStore.java
...p/metadata/repository/typestore/GraphBackedTypeStore.java
+3
-12
DefaultMetadataService.java
...ache/hadoop/metadata/services/DefaultMetadataService.java
+37
-6
GraphPersistenceStrategies.scala
...he/hadoop/metadata/query/GraphPersistenceStrategies.scala
+12
-1
GremlinQuery.scala
...scala/org/apache/hadoop/metadata/query/GremlinQuery.scala
+2
-6
GraphBackedDiscoveryServiceTest.java
...p/metadata/discovery/GraphBackedDiscoveryServiceTest.java
+0
-1
HiveLineageServiceTest.java
...che/hadoop/metadata/discovery/HiveLineageServiceTest.java
+11
-10
GraphBackedMetadataRepositoryTest.java
...a/repository/graph/GraphBackedMetadataRepositoryTest.java
+24
-0
application.properties
repository/src/test/resources/application.properties
+0
-1
application.properties
src/conf/application.properties
+3
-5
pom.xml
webapp/pom.xml
+4
-18
QuickStart.java
.../java/org/apache/hadoop/metadata/examples/QuickStart.java
+8
-8
No files found.
addons/hive-bridge/pom.xml
View file @
80674f0a
...
@@ -35,15 +35,12 @@
...
@@ -35,15 +35,12 @@
<properties>
<properties>
<hive.version>
1.2.0
</hive.version>
<hive.version>
1.2.0
</hive.version>
<calcite.version>
0.9.2-incubating
</calcite.version>
<calcite.version>
0.9.2-incubating
</calcite.version>
<hadoop.version>
2.6.0
</hadoop.version>
</properties>
</properties>
<dependencies>
<dependencies>
<dependency>
<dependency>
<groupId>
org.apache.hadoop
</groupId>
<groupId>
org.apache.hadoop
</groupId>
<artifactId>
hadoop-minikdc
</artifactId>
<artifactId>
hadoop-minikdc
</artifactId>
<version>
${hadoop.version}
</version>
<scope>
test
</scope>
</dependency>
</dependency>
<!-- Logging -->
<!-- Logging -->
...
@@ -71,7 +68,6 @@
...
@@ -71,7 +68,6 @@
<scope>
provided
</scope>
<scope>
provided
</scope>
</dependency>
</dependency>
<dependency>
<dependency>
<groupId>
org.apache.hive
</groupId>
<groupId>
org.apache.hive
</groupId>
<artifactId>
hive-cli
</artifactId>
<artifactId>
hive-cli
</artifactId>
...
@@ -101,13 +97,11 @@
...
@@ -101,13 +97,11 @@
<dependency>
<dependency>
<groupId>
org.apache.hadoop
</groupId>
<groupId>
org.apache.hadoop
</groupId>
<artifactId>
hadoop-client
</artifactId>
<artifactId>
hadoop-client
</artifactId>
<version>
${hadoop.version}
</version>
</dependency>
</dependency>
<dependency>
<dependency>
<groupId>
org.apache.hadoop
</groupId>
<groupId>
org.apache.hadoop
</groupId>
<artifactId>
hadoop-annotations
</artifactId>
<artifactId>
hadoop-annotations
</artifactId>
<version>
${hadoop.version}
</version>
</dependency>
</dependency>
<dependency>
<dependency>
...
@@ -118,7 +112,6 @@
...
@@ -118,7 +112,6 @@
<dependency>
<dependency>
<groupId>
org.apache.hadoop.metadata
</groupId>
<groupId>
org.apache.hadoop.metadata
</groupId>
<artifactId>
metadata-webapp
</artifactId>
<artifactId>
metadata-webapp
</artifactId>
<version>
${project.version}
</version>
<classifier>
classes
</classifier>
<classifier>
classes
</classifier>
</dependency>
</dependency>
...
@@ -127,7 +120,6 @@
...
@@ -127,7 +120,6 @@
<artifactId>
jetty
</artifactId>
<artifactId>
jetty
</artifactId>
<scope>
test
</scope>
<scope>
test
</scope>
</dependency>
</dependency>
</dependencies>
</dependencies>
<build>
<build>
...
@@ -135,7 +127,6 @@
...
@@ -135,7 +127,6 @@
<plugin>
<plugin>
<groupId>
org.apache.maven.plugins
</groupId>
<groupId>
org.apache.maven.plugins
</groupId>
<artifactId>
maven-dependency-plugin
</artifactId>
<artifactId>
maven-dependency-plugin
</artifactId>
<version>
2.10
</version>
<executions>
<executions>
<execution>
<execution>
<id>
copy-bridge-dependencies
</id>
<id>
copy-bridge-dependencies
</id>
...
@@ -311,6 +302,7 @@
...
@@ -311,6 +302,7 @@
<skip>
false
</skip>
<skip>
false
</skip>
</configuration>
</configuration>
</plugin>
</plugin>
<plugin>
<plugin>
<groupId>
org.apache.felix
</groupId>
<groupId>
org.apache.felix
</groupId>
<artifactId>
maven-bundle-plugin
</artifactId>
<artifactId>
maven-bundle-plugin
</artifactId>
...
...
addons/hive-bridge/src/test/java/org/apache/hadoop/metadata/hive/hook/HiveHookIT.java
View file @
80674f0a
...
@@ -77,7 +77,11 @@ public class HiveHookIT {
...
@@ -77,7 +77,11 @@ public class HiveHookIT {
public
void
testCreateDatabase
()
throws
Exception
{
public
void
testCreateDatabase
()
throws
Exception
{
String
dbName
=
"db"
+
random
();
String
dbName
=
"db"
+
random
();
runCommand
(
"create database "
+
dbName
);
runCommand
(
"create database "
+
dbName
);
assertDatabaseIsRegistered
(
dbName
);
//There should be just one entity per dbname
runCommand
(
"drop database "
+
dbName
);
runCommand
(
"create database "
+
dbName
);
assertDatabaseIsRegistered
(
dbName
);
assertDatabaseIsRegistered
(
dbName
);
}
}
...
...
client/pom.xml
View file @
80674f0a
...
@@ -39,9 +39,17 @@
...
@@ -39,9 +39,17 @@
<dependency>
<dependency>
<groupId>
org.apache.hadoop
</groupId>
<groupId>
org.apache.hadoop
</groupId>
<artifactId>
hadoop-common
</artifactId>
</dependency>
<dependency>
<groupId>
org.apache.hadoop
</groupId>
<artifactId>
hadoop-annotations
</artifactId>
</dependency>
<dependency>
<groupId>
org.apache.hadoop
</groupId>
<artifactId>
hadoop-minikdc
</artifactId>
<artifactId>
hadoop-minikdc
</artifactId>
<version>
${hadoop.version}
</version>
<scope>
test
</scope>
</dependency>
</dependency>
<dependency>
<dependency>
...
@@ -65,18 +73,6 @@
...
@@ -65,18 +73,6 @@
</dependency>
</dependency>
<dependency>
<dependency>
<groupId>
org.apache.hadoop
</groupId>
<artifactId>
hadoop-common
</artifactId>
<version>
${hadoop.version}
</version>
</dependency>
<dependency>
<groupId>
org.apache.hadoop
</groupId>
<artifactId>
hadoop-annotations
</artifactId>
<version>
${hadoop.version}
</version>
</dependency>
<dependency>
<groupId>
org.testng
</groupId>
<groupId>
org.testng
</groupId>
<artifactId>
testng
</artifactId>
<artifactId>
testng
</artifactId>
</dependency>
</dependency>
...
@@ -87,7 +83,6 @@
...
@@ -87,7 +83,6 @@
<plugin>
<plugin>
<groupId>
org.apache.maven.plugins
</groupId>
<groupId>
org.apache.maven.plugins
</groupId>
<artifactId>
maven-jar-plugin
</artifactId>
<artifactId>
maven-jar-plugin
</artifactId>
<version>
2.2
</version>
<executions>
<executions>
<execution>
<execution>
<goals>
<goals>
...
@@ -96,6 +91,7 @@
...
@@ -96,6 +91,7 @@
</execution>
</execution>
</executions>
</executions>
</plugin>
</plugin>
<plugin>
<plugin>
<groupId>
org.apache.felix
</groupId>
<groupId>
org.apache.felix
</groupId>
<artifactId>
maven-bundle-plugin
</artifactId>
<artifactId>
maven-bundle-plugin
</artifactId>
...
...
pom.xml
View file @
80674f0a
...
@@ -255,6 +255,39 @@
...
@@ -255,6 +255,39 @@
</exclusions>
</exclusions>
</dependency>
</dependency>
<!-- hadoop -->
<dependency>
<groupId>
org.apache.hadoop
</groupId>
<artifactId>
hadoop-common
</artifactId>
<version>
${hadoop.version}
</version>
</dependency>
<dependency>
<groupId>
org.apache.hadoop
</groupId>
<artifactId>
hadoop-client
</artifactId>
<version>
${hadoop.version}
</version>
</dependency>
<dependency>
<groupId>
org.apache.hadoop
</groupId>
<artifactId>
hadoop-annotations
</artifactId>
<version>
${hadoop.version}
</version>
</dependency>
<dependency>
<groupId>
org.apache.hadoop
</groupId>
<artifactId>
hadoop-hdfs
</artifactId>
<version>
${hadoop.version}
</version>
<scope>
test
</scope>
</dependency>
<dependency>
<groupId>
org.apache.hadoop
</groupId>
<artifactId>
hadoop-minikdc
</artifactId>
<version>
${hadoop.version}
</version>
<scope>
test
</scope>
</dependency>
<!-- commons -->
<!-- commons -->
<dependency>
<dependency>
<groupId>
commons-configuration
</groupId>
<groupId>
commons-configuration
</groupId>
...
@@ -274,6 +307,12 @@
...
@@ -274,6 +307,12 @@
<version>
1.0
</version>
<version>
1.0
</version>
</dependency>
</dependency>
<dependency>
<groupId>
commons-io
</groupId>
<artifactId>
commons-io
</artifactId>
<version>
2.4
</version>
</dependency>
<!-- utilities -->
<!-- utilities -->
<dependency>
<dependency>
<groupId>
com.google.inject
</groupId>
<groupId>
com.google.inject
</groupId>
...
@@ -288,6 +327,18 @@
...
@@ -288,6 +327,18 @@
</dependency>
</dependency>
<dependency>
<dependency>
<groupId>
com.google.inject.extensions
</groupId>
<artifactId>
guice-servlet
</artifactId>
<version>
3.0
</version>
</dependency>
<dependency>
<groupId>
com.sun.jersey.contribs
</groupId>
<artifactId>
jersey-guice
</artifactId>
<version>
1.18.3
</version>
</dependency>
<dependency>
<groupId>
joda-time
</groupId>
<groupId>
joda-time
</groupId>
<artifactId>
joda-time
</artifactId>
<artifactId>
joda-time
</artifactId>
<version>
2.5
</version>
<version>
2.5
</version>
...
@@ -453,10 +504,25 @@
...
@@ -453,10 +504,25 @@
<dependency>
<dependency>
<groupId>
org.apache.hadoop.metadata
</groupId>
<groupId>
org.apache.hadoop.metadata
</groupId>
<artifactId>
metadata-webapp
</artifactId>
<version>
${project.version}
</version>
<classifier>
classes
</classifier>
</dependency>
<dependency>
<groupId>
org.apache.hadoop.metadata
</groupId>
<artifactId>
metadata-client
</artifactId>
<artifactId>
metadata-client
</artifactId>
<version>
${project.version}
</version>
<version>
${project.version}
</version>
</dependency>
</dependency>
<dependency>
<groupId>
org.apache.hadoop.metadata
</groupId>
<artifactId>
metadata-client
</artifactId>
<version>
${project.version}
</version>
<type>
test-jar
</type>
<scope>
test
</scope>
</dependency>
<!--Scala dependencies-->
<!--Scala dependencies-->
<dependency>
<dependency>
<groupId>
org.scala-lang
</groupId>
<groupId>
org.scala-lang
</groupId>
...
@@ -585,6 +651,13 @@
...
@@ -585,6 +651,13 @@
<version>
${fastutil.version}
</version>
<version>
${fastutil.version}
</version>
</dependency>
</dependency>
<!-- supports simple auth handler -->
<dependency>
<groupId>
org.apache.httpcomponents
</groupId>
<artifactId>
httpclient
</artifactId>
<version>
4.2.5
</version>
</dependency>
<!--Test dependencies-->
<!--Test dependencies-->
<dependency>
<dependency>
<groupId>
org.testng
</groupId>
<groupId>
org.testng
</groupId>
...
@@ -695,7 +768,7 @@
...
@@ -695,7 +768,7 @@
<plugin>
<plugin>
<groupId>
org.apache.maven.plugins
</groupId>
<groupId>
org.apache.maven.plugins
</groupId>
<artifactId>
maven-war-plugin
</artifactId>
<artifactId>
maven-war-plugin
</artifactId>
<version>
2.
1.1
</version>
<version>
2.
4
</version>
</plugin>
</plugin>
<plugin>
<plugin>
...
@@ -731,7 +804,7 @@
...
@@ -731,7 +804,7 @@
<plugin>
<plugin>
<groupId>
org.apache.maven.plugins
</groupId>
<groupId>
org.apache.maven.plugins
</groupId>
<artifactId>
maven-dependency-plugin
</artifactId>
<artifactId>
maven-dependency-plugin
</artifactId>
<version>
2.
8
</version>
<version>
2.
10
</version>
</plugin>
</plugin>
<plugin>
<plugin>
...
...
repository/src/main/java/org/apache/hadoop/metadata/GraphTransaction.java
0 → 100644
View file @
80674f0a
/*
* Licensed to the Apache Software Foundation (ASF) under one or more
* contributor license agreements. See the NOTICE file distributed with
* this work for additional information regarding copyright ownership.
* The ASF licenses this file to You under the Apache License, Version 2.0
* (the "License"); you may not use this file except in compliance with
* the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package
org
.
apache
.
hadoop
.
metadata
;
import
java.lang.annotation.ElementType
;
import
java.lang.annotation.Retention
;
import
java.lang.annotation.RetentionPolicy
;
import
java.lang.annotation.Target
;
@Retention
(
RetentionPolicy
.
RUNTIME
)
@Target
(
ElementType
.
METHOD
)
public
@interface
GraphTransaction
{}
repository/src/main/java/org/apache/hadoop/metadata/GraphTransactionInterceptor.java
0 → 100644
View file @
80674f0a
/*
* Licensed to the Apache Software Foundation (ASF) under one or more
* contributor license agreements. See the NOTICE file distributed with
* this work for additional information regarding copyright ownership.
* The ASF licenses this file to You under the Apache License, Version 2.0
* (the "License"); you may not use this file except in compliance with
* the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package
org
.
apache
.
hadoop
.
metadata
;
import
com.google.inject.Inject
;
import
com.thinkaurelius.titan.core.TitanGraph
;
import
org.aopalliance.intercept.MethodInterceptor
;
import
org.aopalliance.intercept.MethodInvocation
;
import
org.apache.hadoop.metadata.repository.graph.GraphProvider
;
import
org.slf4j.Logger
;
import
org.slf4j.LoggerFactory
;
public
class
GraphTransactionInterceptor
implements
MethodInterceptor
{
private
static
final
Logger
LOG
=
LoggerFactory
.
getLogger
(
GraphTransactionInterceptor
.
class
);
private
TitanGraph
titanGraph
;
@Inject
GraphProvider
<
TitanGraph
>
graphProvider
;
public
Object
invoke
(
MethodInvocation
invocation
)
throws
Throwable
{
if
(
titanGraph
==
null
)
{
titanGraph
=
graphProvider
.
get
();
}
try
{
LOG
.
debug
(
"graph rollback to cleanup previous state"
);
titanGraph
.
rollback
();
//cleanup previous state
Object
response
=
invocation
.
proceed
();
titanGraph
.
commit
();
LOG
.
debug
(
"graph commit"
);
return
response
;
}
catch
(
Throwable
t
){
titanGraph
.
rollback
();
LOG
.
debug
(
"graph rollback"
);
throw
t
;
}
}
}
repository/src/main/java/org/apache/hadoop/metadata/RepositoryMetadataModule.java
View file @
80674f0a
...
@@ -19,8 +19,10 @@
...
@@ -19,8 +19,10 @@
package
org
.
apache
.
hadoop
.
metadata
;
package
org
.
apache
.
hadoop
.
metadata
;
import
com.google.inject.Scopes
;
import
com.google.inject.Scopes
;
import
com.google.inject.matcher.Matchers
;
import
com.google.inject.throwingproviders.ThrowingProviderBinder
;
import
com.google.inject.throwingproviders.ThrowingProviderBinder
;
import
com.thinkaurelius.titan.core.TitanGraph
;
import
com.thinkaurelius.titan.core.TitanGraph
;
import
org.aopalliance.intercept.MethodInterceptor
;
import
org.apache.hadoop.metadata.discovery.DiscoveryService
;
import
org.apache.hadoop.metadata.discovery.DiscoveryService
;
import
org.apache.hadoop.metadata.discovery.HiveLineageService
;
import
org.apache.hadoop.metadata.discovery.HiveLineageService
;
import
org.apache.hadoop.metadata.discovery.LineageService
;
import
org.apache.hadoop.metadata.discovery.LineageService
;
...
@@ -40,31 +42,7 @@ import org.apache.hadoop.metadata.services.MetadataService;
...
@@ -40,31 +42,7 @@ import org.apache.hadoop.metadata.services.MetadataService;
* Guice module for Repository module.
* Guice module for Repository module.
*/
*/
public
class
RepositoryMetadataModule
extends
com
.
google
.
inject
.
AbstractModule
{
public
class
RepositoryMetadataModule
extends
com
.
google
.
inject
.
AbstractModule
{
@Override
// Graph Service implementation class
// private Class<? extends GraphService> graphServiceClass;
// MetadataRepositoryService implementation class
private
Class
<?
extends
MetadataRepository
>
metadataRepoClass
;
private
Class
<?
extends
ITypeStore
>
typeStore
;
private
Class
<?
extends
MetadataService
>
metadataService
;
private
Class
<?
extends
DiscoveryService
>
discoveryService
;
private
Class
<?
extends
SearchIndexer
>
searchIndexer
;
private
Class
<?
extends
LineageService
>
lineageService
;
public
RepositoryMetadataModule
()
{
// GraphServiceConfigurator gsp = new GraphServiceConfigurator();
// get the impl classes for the repo and the graph service
// this.graphServiceClass = gsp.getImplClass();
this
.
metadataRepoClass
=
GraphBackedMetadataRepository
.
class
;
this
.
typeStore
=
GraphBackedTypeStore
.
class
;
this
.
metadataService
=
DefaultMetadataService
.
class
;
this
.
discoveryService
=
GraphBackedDiscoveryService
.
class
;
this
.
searchIndexer
=
GraphBackedSearchIndexer
.
class
;
this
.
lineageService
=
HiveLineageService
.
class
;
}
protected
void
configure
()
{
protected
void
configure
()
{
// special wiring for Titan Graph
// special wiring for Titan Graph
ThrowingProviderBinder
.
create
(
binder
())
ThrowingProviderBinder
.
create
(
binder
())
...
@@ -75,22 +53,26 @@ public class RepositoryMetadataModule extends com.google.inject.AbstractModule {
...
@@ -75,22 +53,26 @@ public class RepositoryMetadataModule extends com.google.inject.AbstractModule {
// allow for dynamic binding of the metadata repo & graph service
// allow for dynamic binding of the metadata repo & graph service
// bind the MetadataRepositoryService interface to an implementation
// bind the MetadataRepositoryService interface to an implementation
bind
(
MetadataRepository
.
class
).
to
(
metadataRepoC
lass
);
bind
(
MetadataRepository
.
class
).
to
(
GraphBackedMetadataRepository
.
c
lass
);
// bind the ITypeStore interface to an implementation
// bind the ITypeStore interface to an implementation
bind
(
ITypeStore
.
class
).
to
(
typeStore
);
bind
(
ITypeStore
.
class
).
to
(
GraphBackedTypeStore
.
class
);
// bind the GraphService interface to an implementation
// bind the GraphService interface to an implementation
// bind(GraphService.class).to(graphServiceClass);
// bind(GraphService.class).to(graphServiceClass);
// bind the MetadataService interface to an implementation
// bind the MetadataService interface to an implementation
bind
(
MetadataService
.
class
).
to
(
metadataService
);
bind
(
MetadataService
.
class
).
to
(
DefaultMetadataService
.
class
);
// bind the DiscoveryService interface to an implementation
// bind the DiscoveryService interface to an implementation
bind
(
DiscoveryService
.
class
).
to
(
discoveryService
);
bind
(
DiscoveryService
.
class
).
to
(
GraphBackedDiscoveryService
.
class
);
bind
(
SearchIndexer
.
class
).
to
(
GraphBackedSearchIndexer
.
class
);
bind
(
SearchIndexer
.
class
).
to
(
searchIndexer
);
bind
(
LineageService
.
class
).
to
(
HiveLineageService
.
class
);
bind
(
LineageService
.
class
).
to
(
lineageService
);
MethodInterceptor
interceptor
=
new
GraphTransactionInterceptor
();
requestInjection
(
interceptor
);
bindInterceptor
(
Matchers
.
any
(),
Matchers
.
annotatedWith
(
GraphTransaction
.
class
),
interceptor
);
}
}
}
}
repository/src/main/java/org/apache/hadoop/metadata/discovery/HiveLineageService.java
View file @
80674f0a
...
@@ -20,6 +20,7 @@ package org.apache.hadoop.metadata.discovery;
...
@@ -20,6 +20,7 @@ package org.apache.hadoop.metadata.discovery;
import
com.thinkaurelius.titan.core.TitanGraph
;
import
com.thinkaurelius.titan.core.TitanGraph
;
import
org.apache.commons.configuration.PropertiesConfiguration
;
import
org.apache.commons.configuration.PropertiesConfiguration
;
import
org.apache.hadoop.metadata.GraphTransaction
;
import
org.apache.hadoop.metadata.MetadataException
;
import
org.apache.hadoop.metadata.MetadataException
;
import
org.apache.hadoop.metadata.PropertiesUtil
;
import
org.apache.hadoop.metadata.PropertiesUtil
;
import
org.apache.hadoop.metadata.discovery.graph.DefaultGraphPersistenceStrategy
;
import
org.apache.hadoop.metadata.discovery.graph.DefaultGraphPersistenceStrategy
;
...
@@ -96,6 +97,7 @@ public class HiveLineageService implements LineageService {
...
@@ -96,6 +97,7 @@ public class HiveLineageService implements LineageService {
* @return Lineage Outputs as JSON
* @return Lineage Outputs as JSON
*/
*/
@Override
@Override
@GraphTransaction
public
String
getOutputs
(
String
tableName
)
throws
DiscoveryException
{
public
String
getOutputs
(
String
tableName
)
throws
DiscoveryException
{
LOG
.
info
(
"Fetching lineage outputs for tableName={}"
,
tableName
);
LOG
.
info
(
"Fetching lineage outputs for tableName={}"
,
tableName
);
...
@@ -121,6 +123,7 @@ public class HiveLineageService implements LineageService {
...
@@ -121,6 +123,7 @@ public class HiveLineageService implements LineageService {
* @return Outputs Graph as JSON
* @return Outputs Graph as JSON
*/
*/
@Override
@Override
@GraphTransaction
public
String
getOutputsGraph
(
String
tableName
)
throws
DiscoveryException
{
public
String
getOutputsGraph
(
String
tableName
)
throws
DiscoveryException
{
LOG
.
info
(
"Fetching lineage outputs graph for tableName={}"
,
tableName
);
LOG
.
info
(
"Fetching lineage outputs graph for tableName={}"
,
tableName
);
...
@@ -139,6 +142,7 @@ public class HiveLineageService implements LineageService {
...
@@ -139,6 +142,7 @@ public class HiveLineageService implements LineageService {
* @return Lineage Inputs as JSON
* @return Lineage Inputs as JSON
*/
*/
@Override
@Override
@GraphTransaction
public
String
getInputs
(
String
tableName
)
throws
DiscoveryException
{
public
String
getInputs
(
String
tableName
)
throws
DiscoveryException
{
LOG
.
info
(
"Fetching lineage inputs for tableName={}"
,
tableName
);
LOG
.
info
(
"Fetching lineage inputs for tableName={}"
,
tableName
);
...
@@ -164,6 +168,7 @@ public class HiveLineageService implements LineageService {
...
@@ -164,6 +168,7 @@ public class HiveLineageService implements LineageService {
* @return Inputs Graph as JSON
* @return Inputs Graph as JSON
*/
*/
@Override
@Override
@GraphTransaction
public
String
getInputsGraph
(
String
tableName
)
throws
DiscoveryException
{
public
String
getInputsGraph
(
String
tableName
)
throws
DiscoveryException
{
LOG
.
info
(
"Fetching lineage inputs graph for tableName={}"
,
tableName
);
LOG
.
info
(
"Fetching lineage inputs graph for tableName={}"
,
tableName
);
...
@@ -182,6 +187,7 @@ public class HiveLineageService implements LineageService {
...
@@ -182,6 +187,7 @@ public class HiveLineageService implements LineageService {
* @return Schema as JSON
* @return Schema as JSON
*/
*/
@Override
@Override
@GraphTransaction
public
String
getSchema
(
String
tableName
)
throws
DiscoveryException
{
public
String
getSchema
(
String
tableName
)
throws
DiscoveryException
{
// todo - validate if indeed this is a table type and exists
// todo - validate if indeed this is a table type and exists
String
schemaQuery
=
HIVE_TABLE_TYPE_NAME
String
schemaQuery
=
HIVE_TABLE_TYPE_NAME
...
...
repository/src/main/java/org/apache/hadoop/metadata/discovery/graph/DefaultGraphPersistenceStrategy.java
View file @
80674f0a
...
@@ -22,6 +22,7 @@ import com.thinkaurelius.titan.core.TitanVertex;
...
@@ -22,6 +22,7 @@ import com.thinkaurelius.titan.core.TitanVertex;
import
org.apache.hadoop.metadata.MetadataException
;
import
org.apache.hadoop.metadata.MetadataException
;
import
org.apache.hadoop.metadata.query.Expressions
;
import
org.apache.hadoop.metadata.query.Expressions
;
import
org.apache.hadoop.metadata.query.GraphPersistenceStrategies
;
import
org.apache.hadoop.metadata.query.GraphPersistenceStrategies
;
import
org.apache.hadoop.metadata.query.GraphPersistenceStrategies
$class
;
import
org.apache.hadoop.metadata.query.TypeUtils
;
import
org.apache.hadoop.metadata.query.TypeUtils
;
import
org.apache.hadoop.metadata.repository.MetadataRepository
;
import
org.apache.hadoop.metadata.repository.MetadataRepository
;
import
org.apache.hadoop.metadata.repository.Constants
;
import
org.apache.hadoop.metadata.repository.Constants
;
...
@@ -168,33 +169,12 @@ public class DefaultGraphPersistenceStrategy implements GraphPersistenceStrategi
...
@@ -168,33 +169,12 @@ public class DefaultGraphPersistenceStrategy implements GraphPersistenceStrategi
@Override
@Override
public
String
gremlinCompOp
(
Expressions
.
ComparisonExpression
op
)
{
public
String
gremlinCompOp
(
Expressions
.
ComparisonExpression
op
)
{
switch
(
op
.
symbol
())
{
return
GraphPersistenceStrategies$class
.
gremlinCompOp
(
this
,
op
);
case
"="
:
return
"T.eq"
;
case
"!="
:
return
"T.neq"
;
case
">"
:
return
"T.gt"
;
case
">="
:
return
"T.gte"
;
case
"<"
:
return
"T.lt"
;
case
"<="
:
return
"T.lte"
;
default
:
throw
new
RuntimeException
((
"Comparison operator not supported in Gremlin: "
+
op
));
}
}
}
@Override
@Override
public
String
loopObjectExpression
(
IDataType
<?>
dataType
)
{
public
String
loopObjectExpression
(
IDataType
<?>
dataType
)
{
return
"{it.object."
+
typeAttributeName
()
+
" == '"
+
dataType
.
getName
()
+
"'}"
;
return
GraphPersistenceStrategies$class
.
loopObjectExpression
(
this
,
dataType
)
;
}
}
@Override
@Override
...
@@ -206,4 +186,9 @@ public class DefaultGraphPersistenceStrategy implements GraphPersistenceStrategi
...
@@ -206,4 +186,9 @@ public class DefaultGraphPersistenceStrategy implements GraphPersistenceStrategi
@Override
@Override
public
String
idAttributeName
()
{
return
metadataRepository
.
getIdAttributeName
();
}
public
String
idAttributeName
()
{
return
metadataRepository
.
getIdAttributeName
();
}
@Override
public
String
typeTestExpression
(
String
typeName
)
{
return
GraphPersistenceStrategies$class
.
typeTestExpression
(
this
,
typeName
);
}
}
}
repository/src/main/java/org/apache/hadoop/metadata/discovery/graph/GraphBackedDiscoveryService.java
View file @
80674f0a
...
@@ -25,6 +25,7 @@ import com.thinkaurelius.titan.core.TitanVertex;
...
@@ -25,6 +25,7 @@ import com.thinkaurelius.titan.core.TitanVertex;
import
com.tinkerpop.blueprints.Vertex
;
import
com.tinkerpop.blueprints.Vertex
;
import
com.tinkerpop.gremlin.groovy.Gremlin
;
import
com.tinkerpop.gremlin.groovy.Gremlin
;
import
com.tinkerpop.gremlin.java.GremlinPipeline
;
import
com.tinkerpop.gremlin.java.GremlinPipeline
;
import
org.apache.hadoop.metadata.GraphTransaction
;
import
org.apache.hadoop.metadata.MetadataServiceClient
;
import
org.apache.hadoop.metadata.MetadataServiceClient
;
import
org.apache.hadoop.metadata.discovery.DiscoveryException
;
import
org.apache.hadoop.metadata.discovery.DiscoveryException
;
import
org.apache.hadoop.metadata.discovery.DiscoveryService
;
import
org.apache.hadoop.metadata.discovery.DiscoveryService
;
...
@@ -82,6 +83,7 @@ public class GraphBackedDiscoveryService implements DiscoveryService {
...
@@ -82,6 +83,7 @@ public class GraphBackedDiscoveryService implements DiscoveryService {
//http://www.elastic.co/guide/en/elasticsearch/reference/current/query-dsl-query-string-query
//http://www.elastic.co/guide/en/elasticsearch/reference/current/query-dsl-query-string-query
// .html#query-string-syntax for query syntax
// .html#query-string-syntax for query syntax
@Override
@Override
@GraphTransaction
public
String
searchByFullText
(
String
query
)
throws
DiscoveryException
{
public
String
searchByFullText
(
String
query
)
throws
DiscoveryException
{
String
graphQuery
=
String
.
format
(
"v.%s:(%s)"
,
Constants
.
ENTITY_TEXT_PROPERTY_KEY
,
query
);
String
graphQuery
=
String
.
format
(
"v.%s:(%s)"
,
Constants
.
ENTITY_TEXT_PROPERTY_KEY
,
query
);
LOG
.
debug
(
"Full text query: {}"
,
graphQuery
);
LOG
.
debug
(
"Full text query: {}"
,
graphQuery
);
...
@@ -118,6 +120,7 @@ public class GraphBackedDiscoveryService implements DiscoveryService {
...
@@ -118,6 +120,7 @@ public class GraphBackedDiscoveryService implements DiscoveryService {
* @return JSON representing the type and results.
* @return JSON representing the type and results.
*/
*/
@Override
@Override
@GraphTransaction
public
String
searchByDSL
(
String
dslQuery
)
throws
DiscoveryException
{
public
String
searchByDSL
(
String
dslQuery
)
throws
DiscoveryException
{
LOG
.
info
(
"Executing dsl query={}"
,
dslQuery
);
LOG
.
info
(
"Executing dsl query={}"
,
dslQuery
);
try
{
try
{
...
@@ -155,6 +158,7 @@ public class GraphBackedDiscoveryService implements DiscoveryService {
...
@@ -155,6 +158,7 @@ public class GraphBackedDiscoveryService implements DiscoveryService {
* @throws org.apache.hadoop.metadata.discovery.DiscoveryException
* @throws org.apache.hadoop.metadata.discovery.DiscoveryException
*/
*/
@Override
@Override
@GraphTransaction
public
List
<
Map
<
String
,
String
>>
searchByGremlin
(
String
gremlinQuery
)
public
List
<
Map
<
String
,
String
>>
searchByGremlin
(
String
gremlinQuery
)
throws
DiscoveryException
{
throws
DiscoveryException
{
LOG
.
info
(
"Executing gremlin query={}"
,
gremlinQuery
);
LOG
.
info
(
"Executing gremlin query={}"
,
gremlinQuery
);
...
...
repository/src/main/java/org/apache/hadoop/metadata/repository/graph/GraphBackedMetadataRepository.java
View file @
80674f0a
...
@@ -27,6 +27,7 @@ import com.tinkerpop.blueprints.Edge;
...
@@ -27,6 +27,7 @@ import com.tinkerpop.blueprints.Edge;
import
com.tinkerpop.blueprints.GraphQuery
;
import
com.tinkerpop.blueprints.GraphQuery
;
import
com.tinkerpop.blueprints.Vertex
;
import
com.tinkerpop.blueprints.Vertex
;
import
org.apache.commons.lang.StringUtils
;
import
org.apache.commons.lang.StringUtils
;
import
org.apache.hadoop.metadata.GraphTransaction
;
import
org.apache.hadoop.metadata.MetadataException
;
import
org.apache.hadoop.metadata.MetadataException
;
import
org.apache.hadoop.metadata.repository.Constants
;
import
org.apache.hadoop.metadata.repository.Constants
;
import
org.apache.hadoop.metadata.repository.MetadataRepository
;
import
org.apache.hadoop.metadata.repository.MetadataRepository
;
...
@@ -137,27 +138,23 @@ public class GraphBackedMetadataRepository implements MetadataRepository {
...
@@ -137,27 +138,23 @@ public class GraphBackedMetadataRepository implements MetadataRepository {
}
}
@Override
@Override
@GraphTransaction
public
String
createEntity
(
IReferenceableInstance
typedInstance
)
throws
RepositoryException
{
public
String
createEntity
(
IReferenceableInstance
typedInstance
)
throws
RepositoryException
{
LOG
.
info
(
"adding entity={}"
,
typedInstance
);
LOG
.
info
(
"adding entity={}"
,
typedInstance
);
try
{
try
{
titanGraph
.
rollback
();
final
String
guid
=
instanceToGraphMapper
.
mapTypedInstanceToGraph
(
typedInstance
);
final
String
guid
=
instanceToGraphMapper
.
mapTypedInstanceToGraph
(
typedInstance
);
titanGraph
.
commit
();
// commit if there are no errors
return
guid
;
return
guid
;
}
catch
(
MetadataException
e
)
{
}
catch
(
MetadataException
e
)
{
titanGraph
.
rollback
();
throw
new
RepositoryException
(
e
);
throw
new
RepositoryException
(
e
);
}
}
}
}
@Override
@Override
@GraphTransaction
public
ITypedReferenceableInstance
getEntityDefinition
(
String
guid
)
throws
RepositoryException
{
public
ITypedReferenceableInstance
getEntityDefinition
(
String
guid
)
throws
RepositoryException
{
LOG
.
info
(
"Retrieving entity with guid={}"
,
guid
);
LOG
.
info
(
"Retrieving entity with guid={}"
,
guid
);
try
{
try
{
titanGraph
.
rollback
();
// clean up before starting a query
Vertex
instanceVertex
=
getVertexForGUID
(
guid
);
Vertex
instanceVertex
=
getVertexForGUID
(
guid
);
LOG
.
debug
(
"Found a vertex {} for guid {}"
,
instanceVertex
,
guid
);
LOG
.
debug
(
"Found a vertex {} for guid {}"
,
instanceVertex
,
guid
);
...
@@ -206,9 +203,9 @@ public class GraphBackedMetadataRepository implements MetadataRepository {
...
@@ -206,9 +203,9 @@ public class GraphBackedMetadataRepository implements MetadataRepository {
* @throws RepositoryException
* @throws RepositoryException
*/
*/
@Override
@Override
@GraphTransaction
public
List
<
String
>
getTraitNames
(
String
guid
)
throws
RepositoryException
{
public
List
<
String
>
getTraitNames
(
String
guid
)
throws
RepositoryException
{
LOG
.
info
(
"Retrieving trait names for entity={}"
,
guid
);
LOG
.
info
(
"Retrieving trait names for entity={}"
,
guid
);
titanGraph
.
rollback
();
// clean up before starting a query
Vertex
instanceVertex
=
getVertexForGUID
(
guid
);
Vertex
instanceVertex
=
getVertexForGUID
(
guid
);
return
getTraitNames
(
instanceVertex
);
return
getTraitNames
(
instanceVertex
);
}
}
...
@@ -231,6 +228,7 @@ public class GraphBackedMetadataRepository implements MetadataRepository {
...
@@ -231,6 +228,7 @@ public class GraphBackedMetadataRepository implements MetadataRepository {
* @throws RepositoryException
* @throws RepositoryException
*/
*/
@Override
@Override
@GraphTransaction
public
void
addTrait
(
String
guid
,
public
void
addTrait
(
String
guid
,
ITypedStruct
traitInstance
)
throws
RepositoryException
{
ITypedStruct
traitInstance
)
throws
RepositoryException
{
Preconditions
.
checkNotNull
(
traitInstance
,
"Trait instance cannot be null"
);
Preconditions
.
checkNotNull
(
traitInstance
,
"Trait instance cannot be null"
);
...
@@ -238,22 +236,18 @@ public class GraphBackedMetadataRepository implements MetadataRepository {
...
@@ -238,22 +236,18 @@ public class GraphBackedMetadataRepository implements MetadataRepository {
LOG
.
info
(
"Adding a new trait={} for entity={}"
,
traitName
,
guid
);
LOG
.
info
(
"Adding a new trait={} for entity={}"
,
traitName
,
guid
);
try
{
try
{
titanGraph
.
rollback
();
// clean up before starting a query
Vertex
instanceVertex
=
getVertexForGUID
(
guid
);
Vertex
instanceVertex
=
getVertexForGUID
(
guid
);
// add the trait instance as a new vertex
// add the trait instance as a new vertex
final
String
typeName
=
getTypeName
(
instanceVertex
);
final
String
typeName
=
getTypeName
(
instanceVertex
);
instanceToGraphMapper
.
mapTraitInstanceToVertex
(
traitInstance
,
instanceToGraphMapper
.
mapTraitInstanceToVertex
(
traitInstance
,
getIdFromVertex
(
typeName
,
instanceVertex
),
getIdFromVertex
(
typeName
,
instanceVertex
),
typeName
,
instanceVertex
,
Collections
.<
Id
,
Vertex
>
emptyMap
());
typeName
,
instanceVertex
,
Collections
.<
Id
,
Vertex
>
emptyMap
());
// update the traits in entity once adding trait instance is successful
// update the traits in entity once adding trait instance is successful
((
TitanVertex
)
instanceVertex
)
((
TitanVertex
)
instanceVertex
)
.
addProperty
(
Constants
.
TRAIT_NAMES_PROPERTY_KEY
,
traitName
);
.
addProperty
(
Constants
.
TRAIT_NAMES_PROPERTY_KEY
,
traitName
);
titanGraph
.
commit
();
// commit if there are no errors
}
catch
(
MetadataException
e
)
{
}
catch
(
MetadataException
e
)
{
titanGraph
.
rollback
();
throw
new
RepositoryException
(
e
);
throw
new
RepositoryException
(
e
);
}
}
}
}
...
@@ -266,11 +260,11 @@ public class GraphBackedMetadataRepository implements MetadataRepository {
...
@@ -266,11 +260,11 @@ public class GraphBackedMetadataRepository implements MetadataRepository {
* @throws RepositoryException
* @throws RepositoryException
*/
*/
@Override
@Override
@GraphTransaction
public
void
deleteTrait
(
String
guid
,
String
traitNameToBeDeleted
)
public
void
deleteTrait
(
String
guid
,
String
traitNameToBeDeleted
)
throws
RepositoryException
{
throws
RepositoryException
{
LOG
.
info
(
"Deleting trait={} from entity={}"
,
traitNameToBeDeleted
,
guid
);
LOG
.
info
(
"Deleting trait={} from entity={}"
,
traitNameToBeDeleted
,
guid
);
try
{
try
{
titanGraph
.
rollback
();
// clean up before starting a query
Vertex
instanceVertex
=
getVertexForGUID
(
guid
);
Vertex
instanceVertex
=
getVertexForGUID
(
guid
);
List
<
String
>
traitNames
=
getTraitNames
(
instanceVertex
);
List
<
String
>
traitNames
=
getTraitNames
(
instanceVertex
);
...
@@ -297,11 +291,8 @@ public class GraphBackedMetadataRepository implements MetadataRepository {
...
@@ -297,11 +291,8 @@ public class GraphBackedMetadataRepository implements MetadataRepository {
traitNames
.
remove
(
traitNameToBeDeleted
);
traitNames
.
remove
(
traitNameToBeDeleted
);
updateTraits
(
instanceVertex
,
traitNames
);
updateTraits
(
instanceVertex
,
traitNames
);
}
}
titanGraph
.
commit
();
// commit if there are no errors
}
}
}
catch
(
Exception
e
)
{
}
catch
(
Exception
e
)
{
titanGraph
.
rollback
();
throw
new
RepositoryException
(
e
);
throw
new
RepositoryException
(
e
);
}
}
}
}
...
@@ -318,11 +309,11 @@ public class GraphBackedMetadataRepository implements MetadataRepository {
...
@@ -318,11 +309,11 @@ public class GraphBackedMetadataRepository implements MetadataRepository {
}
}
@Override
@Override
@GraphTransaction
public
void
updateEntity
(
String
guid
,
String
property
,
String
value
)
throws
RepositoryException
{
public
void
updateEntity
(
String
guid
,
String
property
,
String
value
)
throws
RepositoryException
{
LOG
.
info
(
"Adding property {} for entity guid {}"
,
property
,
guid
);
LOG
.
info
(
"Adding property {} for entity guid {}"
,
property
,
guid
);
try
{
try
{
titanGraph
.
rollback
();
// clean up before starting a query
Vertex
instanceVertex
=
GraphHelper
.
findVertexByGUID
(
titanGraph
,
guid
);
Vertex
instanceVertex
=
GraphHelper
.
findVertexByGUID
(
titanGraph
,
guid
);
if
(
instanceVertex
==
null
)
{
if
(
instanceVertex
==
null
)
{
throw
new
RepositoryException
(
"Could not find a vertex for guid "
+
guid
);
throw
new
RepositoryException
(
"Could not find a vertex for guid "
+
guid
);
...
@@ -349,11 +340,8 @@ public class GraphBackedMetadataRepository implements MetadataRepository {
...
@@ -349,11 +340,8 @@ public class GraphBackedMetadataRepository implements MetadataRepository {
instanceToGraphMapper
.
mapAttributesToVertex
(
getIdFromVertex
(
typeName
,
instanceVertex
),
instance
,
instanceToGraphMapper
.
mapAttributesToVertex
(
getIdFromVertex
(
typeName
,
instanceVertex
),
instance
,
instanceVertex
,
new
HashMap
<
Id
,
Vertex
>(),
attributeInfo
,
attributeInfo
.
dataType
());
instanceVertex
,
new
HashMap
<
Id
,
Vertex
>(),
attributeInfo
,
attributeInfo
.
dataType
());
titanGraph
.
commit
();
}
catch
(
Exception
e
)
{
}
catch
(
Exception
e
)
{
throw
new
RepositoryException
(
e
);
throw
new
RepositoryException
(
e
);
}
finally
{
titanGraph
.
rollback
();
}
}
}
}
...
...
repository/src/main/java/org/apache/hadoop/metadata/repository/graph/TitanGraphProvider.java
View file @
80674f0a
...
@@ -18,6 +18,7 @@
...
@@ -18,6 +18,7 @@
package
org
.
apache
.
hadoop
.
metadata
.
repository
.
graph
;
package
org
.
apache
.
hadoop
.
metadata
.
repository
.
graph
;
import
com.google.inject.Provides
;
import
com.thinkaurelius.titan.core.TitanFactory
;
import
com.thinkaurelius.titan.core.TitanFactory
;
import
com.thinkaurelius.titan.core.TitanGraph
;
import
com.thinkaurelius.titan.core.TitanGraph
;
import
org.apache.commons.configuration.Configuration
;
import
org.apache.commons.configuration.Configuration
;
...
@@ -64,6 +65,7 @@ public class TitanGraphProvider implements GraphProvider<TitanGraph> {
...
@@ -64,6 +65,7 @@ public class TitanGraphProvider implements GraphProvider<TitanGraph> {
@Override
@Override
@Singleton
@Singleton
@Provides
public
TitanGraph
get
()
{
public
TitanGraph
get
()
{
Configuration
config
;
Configuration
config
;
try
{
try
{
...
...
repository/src/main/java/org/apache/hadoop/metadata/repository/typestore/GraphBackedTypeStore.java
View file @
80674f0a
...
@@ -25,6 +25,7 @@ import com.tinkerpop.blueprints.Direction;
...
@@ -25,6 +25,7 @@ import com.tinkerpop.blueprints.Direction;
import
com.tinkerpop.blueprints.Edge
;
import
com.tinkerpop.blueprints.Edge
;
import
com.tinkerpop.blueprints.Vertex
;
import
com.tinkerpop.blueprints.Vertex
;
import
org.apache.commons.lang.StringUtils
;
import
org.apache.commons.lang.StringUtils
;
import
org.apache.hadoop.metadata.GraphTransaction
;
import
org.apache.hadoop.metadata.MetadataException
;
import
org.apache.hadoop.metadata.MetadataException
;
import
org.apache.hadoop.metadata.repository.Constants
;
import
org.apache.hadoop.metadata.repository.Constants
;
import
org.apache.hadoop.metadata.repository.graph.GraphProvider
;
import
org.apache.hadoop.metadata.repository.graph.GraphProvider
;
...
@@ -76,10 +77,9 @@ public class GraphBackedTypeStore implements ITypeStore {
...
@@ -76,10 +77,9 @@ public class GraphBackedTypeStore implements ITypeStore {
}
}
@Override
@Override
@GraphTransaction
public
void
store
(
TypeSystem
typeSystem
,
ImmutableList
<
String
>
typeNames
)
throws
MetadataException
{
public
void
store
(
TypeSystem
typeSystem
,
ImmutableList
<
String
>
typeNames
)
throws
MetadataException
{
try
{
ImmutableList
<
String
>
coreTypes
=
typeSystem
.
getCoreTypes
();
ImmutableList
<
String
>
coreTypes
=
typeSystem
.
getCoreTypes
();
titanGraph
.
rollback
();
//Cleanup previous state
for
(
String
typeName
:
typeNames
)
{
for
(
String
typeName
:
typeNames
)
{
if
(!
coreTypes
.
contains
(
typeName
))
{
if
(!
coreTypes
.
contains
(
typeName
))
{
IDataType
dataType
=
typeSystem
.
getDataType
(
IDataType
.
class
,
typeName
);
IDataType
dataType
=
typeSystem
.
getDataType
(
IDataType
.
class
,
typeName
);
...
@@ -107,10 +107,6 @@ public class GraphBackedTypeStore implements ITypeStore {
...
@@ -107,10 +107,6 @@ public class GraphBackedTypeStore implements ITypeStore {
}
}
}
}
}
}
titanGraph
.
commit
();
}
finally
{
titanGraph
.
rollback
();
}
}
}
private
void
storeInGraph
(
EnumType
dataType
)
{
private
void
storeInGraph
(
EnumType
dataType
)
{
...
@@ -206,9 +202,8 @@ public class GraphBackedTypeStore implements ITypeStore {
...
@@ -206,9 +202,8 @@ public class GraphBackedTypeStore implements ITypeStore {
}
}
@Override
@Override
@GraphTransaction
public
TypesDef
restore
()
throws
MetadataException
{
public
TypesDef
restore
()
throws
MetadataException
{
try
{
titanGraph
.
rollback
();
//Cleanup previous state
//Get all vertices for type system
//Get all vertices for type system
Iterator
vertices
=
Iterator
vertices
=
titanGraph
.
query
().
has
(
Constants
.
VERTEX_TYPE_PROPERTY_KEY
,
VERTEX_TYPE
).
vertices
().
iterator
();
titanGraph
.
query
().
has
(
Constants
.
VERTEX_TYPE_PROPERTY_KEY
,
VERTEX_TYPE
).
vertices
().
iterator
();
...
@@ -249,11 +244,7 @@ public class GraphBackedTypeStore implements ITypeStore {
...
@@ -249,11 +244,7 @@ public class GraphBackedTypeStore implements ITypeStore {
throw
new
IllegalArgumentException
(
"Unhandled type category "
+
typeCategory
);
throw
new
IllegalArgumentException
(
"Unhandled type category "
+
typeCategory
);
}
}
}
}
titanGraph
.
commit
();
return
TypeUtils
.
getTypesDef
(
enums
.
build
(),
structs
.
build
(),
traits
.
build
(),
classTypes
.
build
());
return
TypeUtils
.
getTypesDef
(
enums
.
build
(),
structs
.
build
(),
traits
.
build
(),
classTypes
.
build
());
}
finally
{
titanGraph
.
rollback
();
}
}
}
private
EnumTypeDefinition
getEnumType
(
Vertex
vertex
)
{
private
EnumTypeDefinition
getEnumType
(
Vertex
vertex
)
{
...
...
repository/src/main/java/org/apache/hadoop/metadata/services/DefaultMetadataService.java
View file @
80674f0a
...
@@ -26,7 +26,7 @@ import org.apache.commons.lang3.StringEscapeUtils;
...
@@ -26,7 +26,7 @@ import org.apache.commons.lang3.StringEscapeUtils;
import
org.apache.hadoop.metadata.MetadataException
;
import
org.apache.hadoop.metadata.MetadataException
;
import
org.apache.hadoop.metadata.MetadataServiceClient
;
import
org.apache.hadoop.metadata.MetadataServiceClient
;
import
org.apache.hadoop.metadata.ParamChecker
;
import
org.apache.hadoop.metadata.ParamChecker
;
import
org.apache.hadoop.metadata.
RepositoryMetadataModul
e
;
import
org.apache.hadoop.metadata.
classification.InterfaceAudienc
e
;
import
org.apache.hadoop.metadata.discovery.SearchIndexer
;
import
org.apache.hadoop.metadata.discovery.SearchIndexer
;
import
org.apache.hadoop.metadata.listener.EntityChangeListener
;
import
org.apache.hadoop.metadata.listener.EntityChangeListener
;
import
org.apache.hadoop.metadata.listener.TypesChangeListener
;
import
org.apache.hadoop.metadata.listener.TypesChangeListener
;
...
@@ -41,14 +41,20 @@ import org.apache.hadoop.metadata.typesystem.Referenceable;
...
@@ -41,14 +41,20 @@ import org.apache.hadoop.metadata.typesystem.Referenceable;
import
org.apache.hadoop.metadata.typesystem.Struct
;
import
org.apache.hadoop.metadata.typesystem.Struct
;
import
org.apache.hadoop.metadata.typesystem.TypesDef
;
import
org.apache.hadoop.metadata.typesystem.TypesDef
;
import
org.apache.hadoop.metadata.typesystem.json.InstanceSerialization
;
import
org.apache.hadoop.metadata.typesystem.json.InstanceSerialization
;
import
org.apache.hadoop.metadata.typesystem.json.Serialization
$
;
import
org.apache.hadoop.metadata.typesystem.json.TypesSerialization
;
import
org.apache.hadoop.metadata.typesystem.json.TypesSerialization
;
import
org.apache.hadoop.metadata.typesystem.types.*
;
import
org.apache.hadoop.metadata.typesystem.types.AttributeDefinition
;
import
org.apache.hadoop.metadata.typesystem.types.ClassType
;
import
org.apache.hadoop.metadata.typesystem.types.DataTypes
;
import
org.apache.hadoop.metadata.typesystem.types.HierarchicalTypeDefinition
;
import
org.apache.hadoop.metadata.typesystem.types.IDataType
;
import
org.apache.hadoop.metadata.typesystem.types.Multiplicity
;
import
org.apache.hadoop.metadata.typesystem.types.TraitType
;
import
org.apache.hadoop.metadata.typesystem.types.TypeSystem
;
import
org.apache.hadoop.metadata.typesystem.types.utils.TypesUtil
;
import
org.codehaus.jettison.json.JSONException
;
import
org.codehaus.jettison.json.JSONException
;
import
org.codehaus.jettison.json.JSONObject
;
import
org.codehaus.jettison.json.JSONObject
;
import
org.slf4j.Logger
;
import
org.slf4j.Logger
;
import
org.slf4j.LoggerFactory
;
import
org.slf4j.LoggerFactory
;
import
scala.tools.cmd.Meta
;
import
javax.inject.Inject
;
import
javax.inject.Inject
;
import
javax.inject.Singleton
;
import
javax.inject.Singleton
;
...
@@ -94,12 +100,38 @@ public class DefaultMetadataService implements MetadataService {
...
@@ -94,12 +100,38 @@ public class DefaultMetadataService implements MetadataService {
try
{
try
{
TypesDef
typesDef
=
typeStore
.
restore
();
TypesDef
typesDef
=
typeStore
.
restore
();
typeSystem
.
defineTypes
(
typesDef
);
typeSystem
.
defineTypes
(
typesDef
);
createSuperTypes
();
}
catch
(
MetadataException
e
)
{
}
catch
(
MetadataException
e
)
{
throw
new
RuntimeException
(
e
);
throw
new
RuntimeException
(
e
);
}
}
LOG
.
info
(
"Restored type system from the store"
);
LOG
.
info
(
"Restored type system from the store"
);
}
}
private
static
final
AttributeDefinition
NAME_ATTRIBUTE
=
TypesUtil
.
createRequiredAttrDef
(
"name"
,
DataTypes
.
STRING_TYPE
);
private
static
final
AttributeDefinition
DESCRIPTION_ATTRIBUTE
=
TypesUtil
.
createRequiredAttrDef
(
"description"
,
DataTypes
.
STRING_TYPE
);
private
static
final
String
[]
SUPER_TYPES
=
{
"DataSet"
,
"Process"
,
"Infrastructure"
,
};
@InterfaceAudience
.
Private
public
void
createSuperTypes
()
throws
MetadataException
{
if
(
typeSystem
.
isRegistered
(
SUPER_TYPES
[
0
]))
{
return
;
// this is already registered
}
for
(
String
superTypeName
:
SUPER_TYPES
)
{
HierarchicalTypeDefinition
<
ClassType
>
superTypeDefinition
=
TypesUtil
.
createClassTypeDef
(
superTypeName
,
ImmutableList
.<
String
>
of
(),
NAME_ATTRIBUTE
,
DESCRIPTION_ATTRIBUTE
);
typeSystem
.
defineClassType
(
superTypeDefinition
);
}
}
/**
/**
* Creates a new type based on the type system to enable adding
* Creates a new type based on the type system to enable adding
...
@@ -123,10 +155,9 @@ public class DefaultMetadataService implements MetadataService {
...
@@ -123,10 +155,9 @@ public class DefaultMetadataService implements MetadataService {
typeStore
.
store
(
typeSystem
,
ImmutableList
.
copyOf
(
typesAdded
.
keySet
()));
typeStore
.
store
(
typeSystem
,
ImmutableList
.
copyOf
(
typesAdded
.
keySet
()));
onTypesAddedToRepo
(
typesAdded
);
onTypesAddedToRepo
(
typesAdded
);
JSONObject
response
=
new
JSONObject
()
{{
return
new
JSONObject
()
{{
put
(
MetadataServiceClient
.
TYPES
,
typesAdded
.
keySet
());
put
(
MetadataServiceClient
.
TYPES
,
typesAdded
.
keySet
());
}};
}};
return
response
;
}
catch
(
JSONException
e
)
{
}
catch
(
JSONException
e
)
{
LOG
.
error
(
"Unable to create response for types={}"
,
typeDefinition
,
e
);
LOG
.
error
(
"Unable to create response for types={}"
,
typeDefinition
,
e
);
throw
new
MetadataException
(
"Unable to create response"
);
throw
new
MetadataException
(
"Unable to create response"
);
...
...
repository/src/main/scala/org/apache/hadoop/metadata/query/GraphPersistenceStrategies.scala
View file @
80674f0a
...
@@ -112,7 +112,18 @@ trait GraphPersistenceStrategies {
...
@@ -112,7 +112,18 @@ trait GraphPersistenceStrategies {
}
}
def
loopObjectExpression
(
dataType
:
IDataType
[
_
])
=
{
def
loopObjectExpression
(
dataType
:
IDataType
[
_
])
=
{
s
"{it.object.'${typeAttributeName}' == '${dataType.getName}'}"
_typeTestExpression
(
dataType
.
getName
,
"it.object"
)
}
def
typeTestExpression
(
typeName
:
String
)
:
String
=
{
_typeTestExpression
(
typeName
,
"it"
)
}
private
def
_typeTestExpression
(
typeName
:
String
,
itRef
:
String
)
:
String
=
{
s
"""{(${itRef}.'${typeAttributeName}' == '${typeName}') |
|(${itRef}.'${superTypeAttributeName}' ?
|${itRef}.'${superTypeAttributeName}'.contains('${typeName}') : false)}"""
.
stripMargin
.
replace
(
System
.
getProperty
(
"line.separator"
),
""
)
}
}
}
}
...
...
repository/src/main/scala/org/apache/hadoop/metadata/query/GremlinQuery.scala
View file @
80674f0a
...
@@ -185,13 +185,9 @@ class GremlinTranslator(expr: Expression,
...
@@ -185,13 +185,9 @@ class GremlinTranslator(expr: Expression,
private
def
genQuery
(
expr
:
Expression
,
inSelect
:
Boolean
)
:
String
=
expr
match
{
private
def
genQuery
(
expr
:
Expression
,
inSelect
:
Boolean
)
:
String
=
expr
match
{
case
ClassExpression
(
clsName
)
=>
case
ClassExpression
(
clsName
)
=>
val
typeName
=
gPersistenceBehavior
.
typeAttributeName
s
"""filter${gPersistenceBehavior.typeTestExpression(clsName)}"""
val
superTypeName
=
gPersistenceBehavior
.
superTypeAttributeName
s
"""filter{(it.$typeName == "$clsName") | (it.$superTypeName ? it.$superTypeName.contains("$clsName") : false)}"""
case
TraitExpression
(
clsName
)
=>
case
TraitExpression
(
clsName
)
=>
val
typeName
=
gPersistenceBehavior
.
typeAttributeName
s
"""filter${gPersistenceBehavior.typeTestExpression(clsName)}"""
val
superTypeName
=
gPersistenceBehavior
.
superTypeAttributeName
s
"""filter{(it.$typeName == "$clsName") | (it.$superTypeName ? it.$superTypeName.contains("$clsName") : false)}"""
case
fe
@FieldExpression
(
fieldName
,
fInfo
,
child
)
if
fe
.
dataType
.
getTypeCategory
==
TypeCategory
.
PRIMITIVE
=>
{
case
fe
@FieldExpression
(
fieldName
,
fInfo
,
child
)
if
fe
.
dataType
.
getTypeCategory
==
TypeCategory
.
PRIMITIVE
=>
{
val
fN
=
"\""
+
gPersistenceBehavior
.
fieldNameInVertex
(
fInfo
.
dataType
,
fInfo
.
attrInfo
)
+
"\""
val
fN
=
"\""
+
gPersistenceBehavior
.
fieldNameInVertex
(
fInfo
.
dataType
,
fInfo
.
attrInfo
)
+
"\""
child
match
{
child
match
{
...
...
repository/src/test/java/org/apache/hadoop/metadata/discovery/GraphBackedDiscoveryServiceTest.java
View file @
80674f0a
...
@@ -29,7 +29,6 @@ import org.apache.hadoop.metadata.discovery.graph.GraphBackedDiscoveryService;
...
@@ -29,7 +29,6 @@ import org.apache.hadoop.metadata.discovery.graph.GraphBackedDiscoveryService;
import
org.apache.hadoop.metadata.query.HiveTitanSample
;
import
org.apache.hadoop.metadata.query.HiveTitanSample
;
import
org.apache.hadoop.metadata.query.QueryTestsUtils
;
import
org.apache.hadoop.metadata.query.QueryTestsUtils
;
import
org.apache.hadoop.metadata.repository.graph.GraphBackedMetadataRepository
;
import
org.apache.hadoop.metadata.repository.graph.GraphBackedMetadataRepository
;
import
org.apache.hadoop.metadata.repository.graph.GraphBackedSearchIndexer
;
import
org.apache.hadoop.metadata.repository.graph.GraphHelper
;
import
org.apache.hadoop.metadata.repository.graph.GraphHelper
;
import
org.apache.hadoop.metadata.repository.graph.GraphProvider
;
import
org.apache.hadoop.metadata.repository.graph.GraphProvider
;
import
org.apache.hadoop.metadata.typesystem.ITypedReferenceableInstance
;
import
org.apache.hadoop.metadata.typesystem.ITypedReferenceableInstance
;
...
...
repository/src/test/java/org/apache/hadoop/metadata/discovery/HiveLineageServiceTest.java
View file @
80674f0a
...
@@ -57,6 +57,11 @@ import java.util.List;
...
@@ -57,6 +57,11 @@ import java.util.List;
@Guice
(
modules
=
RepositoryMetadataModule
.
class
)
@Guice
(
modules
=
RepositoryMetadataModule
.
class
)
public
class
HiveLineageServiceTest
{
public
class
HiveLineageServiceTest
{
static
{
// this would override super types creation if not first thing
TypeSystem
.
getInstance
().
reset
();
}
@Inject
@Inject
private
DefaultMetadataService
metadataService
;
private
DefaultMetadataService
metadataService
;
...
@@ -71,8 +76,6 @@ public class HiveLineageServiceTest {
...
@@ -71,8 +76,6 @@ public class HiveLineageServiceTest {
@BeforeClass
@BeforeClass
public
void
setUp
()
throws
Exception
{
public
void
setUp
()
throws
Exception
{
TypeSystem
.
getInstance
().
reset
();
setUpTypes
();
setUpTypes
();
setupInstances
();
setupInstances
();
...
@@ -280,9 +283,7 @@ public class HiveLineageServiceTest {
...
@@ -280,9 +283,7 @@ public class HiveLineageServiceTest {
);
);
HierarchicalTypeDefinition
<
ClassType
>
tblClsDef
=
HierarchicalTypeDefinition
<
ClassType
>
tblClsDef
=
TypesUtil
.
createClassTypeDef
(
HIVE_TABLE_TYPE
,
null
,
TypesUtil
.
createClassTypeDef
(
HIVE_TABLE_TYPE
,
ImmutableList
.
of
(
"DataSet"
),
attrDef
(
"name"
,
DataTypes
.
STRING_TYPE
),
attrDef
(
"description"
,
DataTypes
.
STRING_TYPE
),
attrDef
(
"owner"
,
DataTypes
.
STRING_TYPE
),
attrDef
(
"owner"
,
DataTypes
.
STRING_TYPE
),
attrDef
(
"createTime"
,
DataTypes
.
INT_TYPE
),
attrDef
(
"createTime"
,
DataTypes
.
INT_TYPE
),
attrDef
(
"lastAccessTime"
,
DataTypes
.
INT_TYPE
),
attrDef
(
"lastAccessTime"
,
DataTypes
.
INT_TYPE
),
...
@@ -298,8 +299,7 @@ public class HiveLineageServiceTest {
...
@@ -298,8 +299,7 @@ public class HiveLineageServiceTest {
);
);
HierarchicalTypeDefinition
<
ClassType
>
loadProcessClsDef
=
HierarchicalTypeDefinition
<
ClassType
>
loadProcessClsDef
=
TypesUtil
.
createClassTypeDef
(
HIVE_PROCESS_TYPE
,
null
,
TypesUtil
.
createClassTypeDef
(
HIVE_PROCESS_TYPE
,
ImmutableList
.
of
(
"Process"
),
attrDef
(
"name"
,
DataTypes
.
STRING_TYPE
),
attrDef
(
"userName"
,
DataTypes
.
STRING_TYPE
),
attrDef
(
"userName"
,
DataTypes
.
STRING_TYPE
),
attrDef
(
"startTime"
,
DataTypes
.
INT_TYPE
),
attrDef
(
"startTime"
,
DataTypes
.
INT_TYPE
),
attrDef
(
"endTime"
,
DataTypes
.
INT_TYPE
),
attrDef
(
"endTime"
,
DataTypes
.
INT_TYPE
),
...
@@ -401,7 +401,7 @@ public class HiveLineageServiceTest {
...
@@ -401,7 +401,7 @@ public class HiveLineageServiceTest {
"sales fact daily materialized view"
,
"sales fact daily materialized view"
,
reportingDB
,
sd
,
"Joe BI"
,
"Managed"
,
salesFactColumns
,
"Metric"
);
reportingDB
,
sd
,
"Joe BI"
,
"Managed"
,
salesFactColumns
,
"Metric"
);
loadProcess
(
"loadSalesDaily"
,
"John ETL"
,
loadProcess
(
"loadSalesDaily"
,
"
hive query for daily summary"
,
"
John ETL"
,
ImmutableList
.
of
(
salesFact
,
timeDim
),
ImmutableList
.
of
(
salesFactDaily
),
ImmutableList
.
of
(
salesFact
,
timeDim
),
ImmutableList
.
of
(
salesFactDaily
),
"create table as select "
,
"plan"
,
"id"
,
"graph"
,
"create table as select "
,
"plan"
,
"id"
,
"graph"
,
"ETL"
);
"ETL"
);
...
@@ -434,7 +434,7 @@ public class HiveLineageServiceTest {
...
@@ -434,7 +434,7 @@ public class HiveLineageServiceTest {
"sales fact monthly materialized view"
,
"sales fact monthly materialized view"
,
reportingDB
,
sd
,
"Jane BI"
,
"Managed"
,
salesFactColumns
,
"Metric"
);
reportingDB
,
sd
,
"Jane BI"
,
"Managed"
,
salesFactColumns
,
"Metric"
);
loadProcess
(
"loadSalesMonthly"
,
"John ETL"
,
loadProcess
(
"loadSalesMonthly"
,
"
hive query for monthly summary"
,
"
John ETL"
,
ImmutableList
.
of
(
salesFactDaily
),
ImmutableList
.
of
(
salesFactMonthly
),
ImmutableList
.
of
(
salesFactDaily
),
ImmutableList
.
of
(
salesFactMonthly
),
"create table as select "
,
"plan"
,
"id"
,
"graph"
,
"create table as select "
,
"plan"
,
"id"
,
"graph"
,
"ETL"
);
"ETL"
);
...
@@ -496,7 +496,7 @@ public class HiveLineageServiceTest {
...
@@ -496,7 +496,7 @@ public class HiveLineageServiceTest {
return
createInstance
(
referenceable
);
return
createInstance
(
referenceable
);
}
}
Id
loadProcess
(
String
name
,
String
user
,
Id
loadProcess
(
String
name
,
String
description
,
String
user
,
List
<
Id
>
inputTables
,
List
<
Id
>
inputTables
,
List
<
Id
>
outputTables
,
List
<
Id
>
outputTables
,
String
queryText
,
String
queryPlan
,
String
queryText
,
String
queryPlan
,
...
@@ -504,6 +504,7 @@ public class HiveLineageServiceTest {
...
@@ -504,6 +504,7 @@ public class HiveLineageServiceTest {
String
...
traitNames
)
throws
Exception
{
String
...
traitNames
)
throws
Exception
{
Referenceable
referenceable
=
new
Referenceable
(
HIVE_PROCESS_TYPE
,
traitNames
);
Referenceable
referenceable
=
new
Referenceable
(
HIVE_PROCESS_TYPE
,
traitNames
);
referenceable
.
set
(
"name"
,
name
);
referenceable
.
set
(
"name"
,
name
);
referenceable
.
set
(
"description"
,
description
);
referenceable
.
set
(
"user"
,
user
);
referenceable
.
set
(
"user"
,
user
);
referenceable
.
set
(
"startTime"
,
System
.
currentTimeMillis
());
referenceable
.
set
(
"startTime"
,
System
.
currentTimeMillis
());
referenceable
.
set
(
"endTime"
,
System
.
currentTimeMillis
()
+
10000
);
referenceable
.
set
(
"endTime"
,
System
.
currentTimeMillis
()
+
10000
);
...
...
repository/src/test/java/org/apache/hadoop/metadata/repository/graph/GraphBackedMetadataRepositoryTest.java
View file @
80674f0a
...
@@ -379,6 +379,30 @@ public class GraphBackedMetadataRepositoryTest {
...
@@ -379,6 +379,30 @@ public class GraphBackedMetadataRepositoryTest {
Assert
.
assertEquals
(
row
.
getString
(
"name"
),
"Jane"
);
Assert
.
assertEquals
(
row
.
getString
(
"name"
),
"Jane"
);
}
}
@Test
(
dependsOnMethods
=
"testCreateEntity"
)
public
void
testBug37860
()
throws
Exception
{
String
dslQuery
=
"hive_table as t where name = 'bar' "
+
"database where name = 'foo' and description = 'foo database' select t"
;
System
.
out
.
println
(
"Executing dslQuery = "
+
dslQuery
);
String
jsonResults
=
discoveryService
.
searchByDSL
(
dslQuery
);
Assert
.
assertNotNull
(
jsonResults
);
JSONObject
results
=
new
JSONObject
(
jsonResults
);
Assert
.
assertEquals
(
results
.
length
(),
3
);
System
.
out
.
println
(
"results = "
+
results
);
Object
query
=
results
.
get
(
"query"
);
Assert
.
assertNotNull
(
query
);
JSONObject
dataType
=
results
.
getJSONObject
(
"dataType"
);
Assert
.
assertNotNull
(
dataType
);
JSONArray
rows
=
results
.
getJSONArray
(
"rows"
);
Assert
.
assertEquals
(
rows
.
length
(),
1
);
}
/**
/**
* Full text search requires GraphBackedSearchIndexer, and GraphBackedSearchIndexer can't be enabled in
* Full text search requires GraphBackedSearchIndexer, and GraphBackedSearchIndexer can't be enabled in
* GraphBackedDiscoveryServiceTest because of its test data. So, test for full text search is in
* GraphBackedDiscoveryServiceTest because of its test data. So, test for full text search is in
...
...
repository/src/test/resources/application.properties
View file @
80674f0a
...
@@ -27,7 +27,6 @@ metadata.graph.index.search.backend=elasticsearch
...
@@ -27,7 +27,6 @@ metadata.graph.index.search.backend=elasticsearch
metadata.graph.index.search.directory
=
./target/data/es
metadata.graph.index.search.directory
=
./target/data/es
metadata.graph.index.search.elasticsearch.client-only
=
false
metadata.graph.index.search.elasticsearch.client-only
=
false
metadata.graph.index.search.elasticsearch.local-mode
=
true
metadata.graph.index.search.elasticsearch.local-mode
=
true
metadata.graph.index.search.elasticsearch.create.sleep
=
1000
######### Hive Lineage Configs #########
######### Hive Lineage Configs #########
...
...
src/conf/application.properties
View file @
80674f0a
...
@@ -29,14 +29,12 @@ metadata.graph.index.search.elasticsearch.local-mode=true
...
@@ -29,14 +29,12 @@ metadata.graph.index.search.elasticsearch.local-mode=true
metadata.graph.index.search.elasticsearch.create.sleep
=
2000
metadata.graph.index.search.elasticsearch.create.sleep
=
2000
######### Hive Lineage Configs #########
######### Hive Lineage Configs #########
# This models
follows the quick-start guide
# This models
reflects the base super types for Data and Process
metadata.lineage.hive.table.type.name
=
Table
metadata.lineage.hive.table.type.name
=
DataSet
metadata.lineage.hive.table.column.name
=
columns
metadata.lineage.hive.table.column.name
=
columns
metadata.lineage.hive.process.type.name
=
Load
Process
metadata.lineage.hive.process.type.name
=
Process
metadata.lineage.hive.process.inputs.name
=
inputTables
metadata.lineage.hive.process.inputs.name
=
inputTables
metadata.lineage.hive.process.outputs.name
=
outputTables
metadata.lineage.hive.process.outputs.name
=
outputTables
#Currently unused
#metadata.lineage.hive.column.type.name=Column
######### Security Properties #########
######### Security Properties #########
...
...
webapp/pom.xml
View file @
80674f0a
...
@@ -48,44 +48,31 @@
...
@@ -48,44 +48,31 @@
<artifactId>
metadata-repository
</artifactId>
<artifactId>
metadata-repository
</artifactId>
</dependency>
</dependency>
<!--<dependency>-->
<!--<groupId>org.apache.hadoop.metadata</groupId>-->
<!--<artifactId>metadata-client</artifactId>-->
<!--</dependency>-->
<dependency>
<dependency>
<groupId>
org.apache.hadoop.metadata
</groupId>
<groupId>
org.apache.hadoop.metadata
</groupId>
<artifactId>
metadata-client
</artifactId>
<artifactId>
metadata-client
</artifactId>
<version>
${version}
</version>
<type>
test-jar
</type>
<type>
test-jar
</type>
<scope>
test
</scope>
</dependency>
</dependency>
<dependency>
<dependency>
<groupId>
org.apache.hadoop
</groupId>
<groupId>
org.apache.hadoop
</groupId>
<artifactId>
hadoop-common
</artifactId>
<artifactId>
hadoop-common
</artifactId>
<version>
${hadoop.version}
</version>
</dependency>
</dependency>
<dependency>
<dependency>
<groupId>
org.apache.hadoop
</groupId>
<groupId>
org.apache.hadoop
</groupId>
<artifactId>
hadoop-minikdc
</artifactId>
<artifactId>
hadoop-minikdc
</artifactId>
<version>
${hadoop.version}
</version>
<scope>
test
</scope>
</dependency>
</dependency>
<dependency>
<dependency>
<groupId>
org.apache.hadoop
</groupId>
<groupId>
org.apache.hadoop
</groupId>
<artifactId>
hadoop-hdfs
</artifactId>
<artifactId>
hadoop-hdfs
</artifactId>
<version>
${hadoop.version}
</version>
<scope>
test
</scope>
</dependency>
</dependency>
<!-- supports simple auth handler -->
<!-- supports simple auth handler -->
<dependency>
<dependency>
<groupId>
org.apache.httpcomponents
</groupId>
<groupId>
org.apache.httpcomponents
</groupId>
<artifactId>
httpclient
</artifactId>
<artifactId>
httpclient
</artifactId>
<version>
4.2.5
</version>
</dependency>
</dependency>
<dependency>
<dependency>
...
@@ -171,19 +158,16 @@
...
@@ -171,19 +158,16 @@
<dependency>
<dependency>
<groupId>
com.google.inject.extensions
</groupId>
<groupId>
com.google.inject.extensions
</groupId>
<artifactId>
guice-servlet
</artifactId>
<artifactId>
guice-servlet
</artifactId>
<version>
3.0
</version>
</dependency>
</dependency>
<dependency>
<dependency>
<groupId>
com.google.code.gson
</groupId>
<groupId>
com.google.code.gson
</groupId>
<artifactId>
gson
</artifactId>
<artifactId>
gson
</artifactId>
<version>
2.3.1
</version>
</dependency>
</dependency>
<dependency>
<dependency>
<groupId>
com.sun.jersey.contribs
</groupId>
<groupId>
com.sun.jersey.contribs
</groupId>
<artifactId>
jersey-guice
</artifactId>
<artifactId>
jersey-guice
</artifactId>
<version>
1.18.3
</version>
</dependency>
</dependency>
<dependency>
<dependency>
...
@@ -194,7 +178,6 @@
...
@@ -194,7 +178,6 @@
<dependency>
<dependency>
<groupId>
commons-io
</groupId>
<groupId>
commons-io
</groupId>
<artifactId>
commons-io
</artifactId>
<artifactId>
commons-io
</artifactId>
<version>
2.4
</version>
</dependency>
</dependency>
</dependencies>
</dependencies>
...
@@ -225,7 +208,6 @@
...
@@ -225,7 +208,6 @@
<plugin>
<plugin>
<groupId>
org.apache.maven.plugins
</groupId>
<groupId>
org.apache.maven.plugins
</groupId>
<artifactId>
maven-war-plugin
</artifactId>
<artifactId>
maven-war-plugin
</artifactId>
<version>
2.4
</version>
<configuration>
<configuration>
<attachClasses>
true
</attachClasses>
<attachClasses>
true
</attachClasses>
<webResources>
<webResources>
...
@@ -362,6 +344,10 @@
...
@@ -362,6 +344,10 @@
<inherited>
true
</inherited>
<inherited>
true
</inherited>
<extensions>
true
</extensions>
<extensions>
true
</extensions>
</plugin>
</plugin>
<plugin>
<groupId>
net.alchim31.maven
</groupId>
<artifactId>
scala-maven-plugin
</artifactId>
</plugin>
</plugins>
</plugins>
</build>
</build>
</project>
</project>
webapp/src/main/java/org/apache/hadoop/metadata/examples/QuickStart.java
View file @
80674f0a
...
@@ -128,9 +128,7 @@ public class QuickStart {
...
@@ -128,9 +128,7 @@ public class QuickStart {
);
);
HierarchicalTypeDefinition
<
ClassType
>
tblClsDef
=
HierarchicalTypeDefinition
<
ClassType
>
tblClsDef
=
TypesUtil
.
createClassTypeDef
(
TABLE_TYPE
,
null
,
TypesUtil
.
createClassTypeDef
(
TABLE_TYPE
,
ImmutableList
.
of
(
"DataSet"
),
attrDef
(
"name"
,
DataTypes
.
STRING_TYPE
),
attrDef
(
"description"
,
DataTypes
.
STRING_TYPE
),
new
AttributeDefinition
(
"db"
,
DATABASE_TYPE
,
new
AttributeDefinition
(
"db"
,
DATABASE_TYPE
,
Multiplicity
.
REQUIRED
,
false
,
null
),
Multiplicity
.
REQUIRED
,
false
,
null
),
new
AttributeDefinition
(
"sd"
,
STORAGE_DESC_TYPE
,
new
AttributeDefinition
(
"sd"
,
STORAGE_DESC_TYPE
,
...
@@ -149,8 +147,7 @@ public class QuickStart {
...
@@ -149,8 +147,7 @@ public class QuickStart {
);
);
HierarchicalTypeDefinition
<
ClassType
>
loadProcessClsDef
=
HierarchicalTypeDefinition
<
ClassType
>
loadProcessClsDef
=
TypesUtil
.
createClassTypeDef
(
LOAD_PROCESS_TYPE
,
null
,
TypesUtil
.
createClassTypeDef
(
LOAD_PROCESS_TYPE
,
ImmutableList
.
of
(
"Process"
),
attrDef
(
"name"
,
DataTypes
.
STRING_TYPE
),
attrDef
(
"userName"
,
DataTypes
.
STRING_TYPE
),
attrDef
(
"userName"
,
DataTypes
.
STRING_TYPE
),
attrDef
(
"startTime"
,
DataTypes
.
INT_TYPE
),
attrDef
(
"startTime"
,
DataTypes
.
INT_TYPE
),
attrDef
(
"endTime"
,
DataTypes
.
INT_TYPE
),
attrDef
(
"endTime"
,
DataTypes
.
INT_TYPE
),
...
@@ -273,7 +270,7 @@ public class QuickStart {
...
@@ -273,7 +270,7 @@ public class QuickStart {
"sales fact daily materialized view"
,
reportingDB
,
sd
,
"sales fact daily materialized view"
,
reportingDB
,
sd
,
"Joe BI"
,
"Managed"
,
salesFactColumns
,
"Metric"
);
"Joe BI"
,
"Managed"
,
salesFactColumns
,
"Metric"
);
loadProcess
(
"loadSalesDaily"
,
"John ETL"
,
loadProcess
(
"loadSalesDaily"
,
"
hive query for daily summary"
,
"
John ETL"
,
ImmutableList
.
of
(
salesFact
,
timeDim
),
ImmutableList
.
of
(
salesFactDaily
),
ImmutableList
.
of
(
salesFact
,
timeDim
),
ImmutableList
.
of
(
salesFactDaily
),
"create table as select "
,
"plan"
,
"id"
,
"graph"
,
"create table as select "
,
"plan"
,
"id"
,
"graph"
,
"ETL"
);
"ETL"
);
...
@@ -288,7 +285,7 @@ public class QuickStart {
...
@@ -288,7 +285,7 @@ public class QuickStart {
"sales fact monthly materialized view"
,
"sales fact monthly materialized view"
,
reportingDB
,
sd
,
"Jane BI"
,
"Managed"
,
salesFactColumns
,
"Metric"
);
reportingDB
,
sd
,
"Jane BI"
,
"Managed"
,
salesFactColumns
,
"Metric"
);
loadProcess
(
"loadSalesMonthly"
,
"John ETL"
,
loadProcess
(
"loadSalesMonthly"
,
"
hive query for monthly summary"
,
"
John ETL"
,
ImmutableList
.
of
(
salesFactDaily
),
ImmutableList
.
of
(
salesFactMonthly
),
ImmutableList
.
of
(
salesFactDaily
),
ImmutableList
.
of
(
salesFactMonthly
),
"create table as select "
,
"plan"
,
"id"
,
"graph"
,
"create table as select "
,
"plan"
,
"id"
,
"graph"
,
"ETL"
);
"ETL"
);
...
@@ -362,7 +359,7 @@ public class QuickStart {
...
@@ -362,7 +359,7 @@ public class QuickStart {
return
createInstance
(
referenceable
);
return
createInstance
(
referenceable
);
}
}
Id
loadProcess
(
String
name
,
String
user
,
Id
loadProcess
(
String
name
,
String
description
,
String
user
,
List
<
Id
>
inputTables
,
List
<
Id
>
inputTables
,
List
<
Id
>
outputTables
,
List
<
Id
>
outputTables
,
String
queryText
,
String
queryPlan
,
String
queryText
,
String
queryPlan
,
...
@@ -370,6 +367,7 @@ public class QuickStart {
...
@@ -370,6 +367,7 @@ public class QuickStart {
String
...
traitNames
)
throws
Exception
{
String
...
traitNames
)
throws
Exception
{
Referenceable
referenceable
=
new
Referenceable
(
LOAD_PROCESS_TYPE
,
traitNames
);
Referenceable
referenceable
=
new
Referenceable
(
LOAD_PROCESS_TYPE
,
traitNames
);
referenceable
.
set
(
"name"
,
name
);
referenceable
.
set
(
"name"
,
name
);
referenceable
.
set
(
"description"
,
description
);
referenceable
.
set
(
"user"
,
user
);
referenceable
.
set
(
"user"
,
user
);
referenceable
.
set
(
"startTime"
,
System
.
currentTimeMillis
());
referenceable
.
set
(
"startTime"
,
System
.
currentTimeMillis
());
referenceable
.
set
(
"endTime"
,
System
.
currentTimeMillis
()
+
10000
);
referenceable
.
set
(
"endTime"
,
System
.
currentTimeMillis
()
+
10000
);
...
@@ -465,6 +463,8 @@ public class QuickStart {
...
@@ -465,6 +463,8 @@ public class QuickStart {
*/
*/
"Table where name=\"sales_fact\", columns"
,
"Table where name=\"sales_fact\", columns"
,
"Table where name=\"sales_fact\", columns as column select column.name, column.dataType, column.comment"
,
"Table where name=\"sales_fact\", columns as column select column.name, column.dataType, column.comment"
,
"from DataSet"
,
"from Process"
,
};
};
}
}
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment