mammoth-fountain-32989
04/22/2022, 3:31 PMmodern-monitor-81461
04/22/2022, 4:32 PMdisable_dbt_node_creation
to true
and once with false
) and we feel like having the DBT nodes is better since it tells the user how the resulting dataset was created. But I can totally see that the lineage can become messy once DBT is used all over the place in your pipelines and I think I recall that somewhere on DataHub's roadmap, there was a planned feature where you could hide elements from the lineage and that could be used for showing or not DBT intermediate nodes...
I tried to use the meta_mapping
feature of the DBT source to create a Glossary Term depending on a meta value (our DBT model has a meta field describing the data tier, Bronze/Silver/Gold), but I want to have this term applied to the "platform dataset" and not only to the DBT node... I don't want to have to go in DataHub and manually add those terms to the resulting dataset, I want everything being driven by DBT and the source model. What I would like to know is how are you guys using DBT with DataHub? Is what I'm asking for something impossible or silly? Looking for advice on how to best leverage DBT with DataHub...
Thanks!fierce-city-89572
04/22/2022, 5:32 PMsome-minister-22606
04/23/2022, 3:23 PMblue-hair-87908
04/24/2022, 12:55 AMalert-football-80212
04/24/2022, 9:51 AMstraight-telephone-84434
04/25/2022, 2:16 PMmodern-belgium-81337
04/25/2022, 11:37 PM./gradlew build
To honour the JVM settings for this build a single-use Daemon process will be forked. See <https://docs.gradle.org/6.9.2/userguide/gradle_daemon.html#sec:disabling_the_daemon>.
Daemon will be stopped at the end of the build
FAILURE: Build failed with an exception.
* Where:
Settings file '/Users/thnguyen/dev/rivian/datahub/datahub/settings.gradle'
* What went wrong:
Could not compile settings file '/Users/thnguyen/dev/rivian/datahub/datahub/settings.gradle'.
> startup failed:
General error during semantic analysis: Unsupported class file major version 61
java.lang.IllegalArgumentException: Unsupported class file major version 61
I followed every step in the documentation here https://datahubproject.io/docs/developers/
Is there anything that I’m missing?cool-architect-34612
04/26/2022, 2:36 AM'docker pull acryldata/datahub-upgrade:head && docker run --env-file docker_env/datahub-upgrade.env acryldata/datahub-upgrade:head -u RestoreIndices'
how can I solve this?
Starting upgrade with id RestoreIndices...
Cleanup has not been requested.
Skipping Step 1/3: ClearSearchServiceStep...
Cleanup has not been requested.
Skipping Step 2/3: ClearGraphServiceStep...
Executing Step 3/3: SendMAEStep...
Sending MAE from local DB...
Found 38084 latest aspects in aspects table
Reading rows 0 through 1000 from the aspects table.
2022-04-26 02:34:24.372 WARN 1 --- [ad | producer-1] org.apache.kafka.clients.NetworkClient : [Producer clientId=producer-1] Connection to node -1 (localhost/127.0.0.1:9092) could not be established. Broker may not be available.
2022-04-26 02:34:24.372 WARN 1 --- [ad | producer-1] org.apache.kafka.clients.NetworkClient : [Producer clientId=producer-1] Bootstrap broker localhost:9092 (id: -1 rack: null) disconnected
2022-04-26 02:34:25.577 WARN 1 --- [ad | producer-1] org.apache.kafka.clients.NetworkClient : [Producer clientId=producer-1] Connection to node -1 (localhost/127.0.0.1:9092) could not be established. Broker may not be available.
2022-04-26 02:34:25.577 WARN 1 --- [ad | producer-1] org.apache.kafka.clients.NetworkClient : [Producer clientId=producer-1] Bootstrap broker localhost:9092 (id: -1 rack: null) disconnected
2022-04-26 02:34:26.481 WARN 1 --- [ad | producer-1] org.apache.kafka.clients.NetworkClient : [Producer clientId=producer-1] Connection to node -1 (localhost/127.0.0.1:9092) could not be established. Broker may not be available.
2022-04-26 02:34:26.481 WARN 1 --- [ad | producer-1] org.apache.kafka.clients.NetworkClient : [Producer clientId=producer-1] Bootstrap broker localhost:9092 (id: -1 rack: null) disconnected
2022-04-26 02:34:27.585 WARN 1 --- [ad | producer-1] org.apache.kafka.clients.NetworkClient : [Producer clientId=producer-1] Connection to node -1 (localhost/127.0.0.1:9092) could not be established. Broker may not be available.
2022-04-26 02:34:27.586 WARN 1 --- [ad | producer-1] org.apache.kafka.clients.NetworkClient : [Producer clientId=producer-1] Bootstrap broker localhost:9092 (id: -1 rack: null) disconnected
mammoth-fountain-32989
04/26/2022, 9:32 AMlemon-terabyte-66903
04/26/2022, 3:06 PMpowerful-easter-94567
04/27/2022, 8:02 AMbitter-dusk-52400
04/28/2022, 7:47 AMastonishing-guitar-79208
04/28/2022, 9:17 AMrhythmic-beard-86717
04/28/2022, 12:44 PMcurved-football-28924
04/28/2022, 4:48 PMproud-chef-44773
04/29/2022, 1:04 PMmany-guitar-67205
05/02/2022, 8:18 AMPolicies can be managed under the /policies page, or accessed inside the Control Center, a slide-out menu appearing on the left side of the DataHub UI1. What is the Control Center and this slide-out menu? Or is that documentation out of date? 2. can you use regex in the condition? the JSON has a
"condition": "EQUALS"
hinting at other options, but the code only seems to support EQUALS
bitter-dusk-52400
05/02/2022, 8:55 AMpowerful-librarian-82760
05/02/2022, 3:42 PMorange-tailor-45265
05/03/2022, 2:32 AMnamespace com.linkedin.schema
/**
* Schema holder for Thrift schema types.
*/
record ThriftSchema {
/**
* The native schema in the dataset's platform.
*/
rawSchema: string
fields: array[ThriftField]
annotations: optional array[Annotation]
namespace_: optional map[string, string]
}
I expected that I can generate the python code from this by excuting ./gradle :metadata-ingestion:codegen
and I expected the generated python code appearing in this path: datahub/metadata-ingestion/src/datahub/metadata/com/linkedin/pegasus2avro/schema/__init__.py
as well as /home/waylee/datahub/metadata-ingestion/src/datahub/metadata/schema_classes.py
because OtherSchema and many other pythohn codes generated by pdl file appear here. My ThriftSchema,pdl located in datahub/metadata-models/src/main/pegasus/com/linkedin/schema/ThriftSchema.pdl
.
Is there anything wrong in my process? My context is creating a schema to ingest thrift file.bitter-dusk-52400
05/03/2022, 7:11 AM'{"query" : "mutation createDomain { createDomain(input: { name:\"det_test\",description: \"from graphql api for test\"})}"}'
if i execute the above query on the curl command i can able to create domains but i tried using python packages like gql i got error like “__init__() got an unexpected keyword argument ‘allowed_methods’”
please refer below python code:
from gql import Client, gql
from gql.transport.requests import RequestsHTTPTransport
transport = RequestsHTTPTransport(
url='<http://localhost:8080/api/graphql>',
headers= {
'X-DataHub-Actor': 'urn:li:corpuser:datahub',
'Content-Type': 'application/json'
},
verify=True,
retries=3,
)
client = Client(transport=transport, fetch_schema_from_transport=True)
query = gql(
'''
mutation createDomain
{ createDomain(input:
{ name:"det_test1",description: "from graphql api for test"})
}
'''
)
result = client.execute(query)
print(result)
bright-furniture-98709
05/03/2022, 2:03 PMbored-dress-52175
05/03/2022, 2:11 PMbored-dress-52175
05/03/2022, 8:03 PMwonderful-egg-79350
05/04/2022, 3:10 AMcurved-crayon-1929
05/04/2022, 7:50 AMfresh-napkin-5247
05/04/2022, 12:22 PMshy-kitchen-7972
05/04/2022, 1:03 PMable-optician-93924
05/04/2022, 4:10 PM