square-solstice-69079
05/12/2022, 6:10 AMmost-plumber-32123
05/12/2022, 6:42 AM[2022-05-12 12:11:31,452] INFO {datahub.cli.ingest_cli:96} - DataHub CLI version: 0.8.34.1
[2022-05-12 12:11:31,738] ERROR {datahub.entrypoints:165} - Unable to connect to <http://localhost:9002/api/gms/config> with status_code: 401. Maybe you need to set up authentication? Please check your configuration and make sure you are talking to the DataHub GMS (usually <datahub-gms-host>:8080) or Frontend GMS API (usually <frontend>:9002/api/gms).
[2022-05-12 12:11:31,738] INFO {datahub.entrypoints:176} - DataHub CLI version: 0.8.34.1 at C:\Users\*****\AppData\Local\Programs\Python\Python39\lib\site-packages\datahub\__init__.py
[2022-05-12 12:11:31,738] INFO {datahub.entrypoints:179} - Python version: 3.9.7 (tags/v3.9.7:1016ef3, Aug 30 2021, 20:19:38) [MSC v.1929 64 bit (AMD64)] at C:\Users\*****\AppData\Local\Programs\Python\Python39\python.exe on Windows-10-10.0.22000-SP0
[2022-05-12 12:11:31,738] INFO {datahub.entrypoints:182} - GMS config {}
handsome-football-66174
05/12/2022, 6:12 PMaloof-author-52810
05/12/2022, 8:08 PMmodern-zoo-97059
05/13/2022, 2:47 AMdatahub_usage_event
created in ElasticSearch? It's missing.sticky-dawn-95000
05/13/2022, 5:42 AMminiature-journalist-76345
05/13/2022, 3:59 PMDuplicate entry 'urn:li:glossaryTerm:test_term-glossaryTermKey-0' for key 'PRIMARY'
Had the same error with platforms few months ago. More information in the threadhandsome-football-66174
05/13/2022, 5:13 PMshy-parrot-64120
05/14/2022, 8:46 PMCaused by: org.springframework.beans.factory.BeanCreationException: Error creating bean with name 'cassandraSession' defined in class path resource [org/springframework/boot/autoconfigure/cassandra/CassandraAutoConfiguration.class]: Bean instantiation via factory method failed; nested exception is org.springframework.beans.BeanInstantiationException: Failed to instantiate [com.datastax.oss.driver.api.core.CqlSession]: Factory method 'cassandraSession' threw exception; nested exception is com.datastax.oss.driver.api.core.AllNodesFailedException: Could not reach any contact point, make sure you've provided valid addresses (showing first 1 nodes, use getAllErrors() for more): Node(endPoint=/127.0.0.1:9042, hostId=null, hashCode=41fdd1a):
[com.datastax.oss.driver.api.core.connection.ConnectionInitException: [s0|control|connecting...] Protocol initialization request, step 1 (OPTIONS): failed to send request (io.netty.channel.StacklessClosedChannelException)]
any suggestions why so? We are not mentioning Cassandra anywhere in setupswift-breakfast-25077
05/16/2022, 9:58 AM- name: datahub_action
action:
module_name: datahub.integrations.great_expectations.action
class_name: DataHubValidationAction
server_url: <http://localhost:8080> #datahub server url
Getting this message when checkpoint runs :icy-portugal-26250
05/16/2022, 11:01 AMdocker/quickstart.sh
(using m1 compose-files). The datahub-gms
container is unhealty and checkking the logs it cannot connect to the elasticsearch service:
Problem with request: Get "<http://elasticsearch:9200>": dial tcp 172.30.0.2:9200: connect: connection refused. Sleeping 1s
Is there a way to troubleshoot this?stale-exabyte-65991
05/16/2022, 2:08 PMbrash-fountain-36115
05/16/2022, 2:13 PM'[2022-05-13 14:18:26,848] WARNING {datahub.ingestion.source.sql.snowflake:496} - lineage => Extracting lineage from Snowflake '
'failed.Please check your premissions. Continuing...\n'
'Error was (snowflake.connector.errors.ProgrammingError) 003030 (02000): SQL compilation error:\n'
'Shared database is no longer available for use. It will need to be re-created if and when the publisher makes it available again.\n'
'[SQL: \n'
'WITH table_lineage_history AS (\n'
' SELECT\n'
' r.value:"objectName" AS upstream_table_name,\n'
' r.value:"objectDomain" AS upstream_table_domain,\n'
' r.value:"columns" AS upstream_table_columns,\n'
' w.value:"objectName" AS downstream_table_name,\n'
' w.value:"objectDomain" AS downstream_table_domain,\n'
' w.value:"columns" AS downstream_table_columns,\n'
' t.query_start_time AS query_start_time\n'
' FROM\n'
' (SELECT * from snowflake.account_usage.access_history) t,\n'
' lateral flatten(input => t.DIRECT_OBJECTS_ACCESSED) r,\n'
' lateral flatten(input => t.OBJECTS_MODIFIED) w\n'
' WHERE r.value:"objectId" IS NOT NULL\n'
' AND w.value:"objectId" IS NOT NULL\n'
' AND w.value:"objectName" NOT LIKE \'%.GE_TMP_%\'\n'
' AND w.value:"objectName" NOT LIKE \'%.GE_TEMP_%\'\n'
' AND t.query_start_time >= to_timestamp_ltz(0, 3)\n'
' AND t.query_start_time < to_timestamp_ltz(1652486400000, 3))\n'
'SELECT upstream_table_name, downstream_table_name, upstream_table_columns, downstream_table_columns\n'
'FROM table_lineage_history\n'
"WHERE upstream_table_domain in ('Table', 'External table') and downstream_table_domain = 'Table'\n"
'QUALIFY ROW_NUMBER() OVER (PARTITION BY downstream_table_name, upstream_table_name ORDER BY query_start_time DESC) = 1 ]\n'
'(Background on this error at: <http://sqlalche.me/e/13/f405>).\n'
'[2022-05-13 14:18:26,848] INFO {datahub.ingestion.source.sql.snowflake:449} - A total of 0 Table->Table edges found for 0 downstream '
'tables.\n'
Any hints what I can check to identify the root cause?red-pizza-28006
05/16/2022, 3:28 PMself = <botocore.client.S3 object at 0x7fb305a355b0>
operation_name = 'GetObject'
api_params = {'Bucket': 'datalake-prod-bqtos3',
'Key': 'script.py'}
http.status_code = 403
error_code = 'AccessDenied'
error_class = <class 'botocore.exceptions.ClientError'>
self.exceptions.from_code = <method 'BaseClientExceptions.from_code' of <botocore.errorfactory.S3Exceptions object at 0x7fb305b1f580> errorfactory.p
y:30>
parsed_response = {'Error': {'Code': 'AccessDenied',
'Message': 'Access Denied'},
'ResponseMetadata': {'RequestId': 'BNJ2Z2VWHBE64YYV',
'HostId': 'fwy92BEUGB+HlSJufrCUKRxa2WZ877BXQNdqWxX5Tx7WR7Br+6bCy16bv7GFTU1ICR0oJh4ingg=',
'HTTPStatusCode': 403,
'HTTPHeaders': {...},
'RetryAttempts': 0}}
My AWS user has full access to Glue, what additional access do i need here to be able to read the glue catalog?late-country-26504
05/16/2022, 6:26 PMmillions-waiter-49836
05/16/2022, 10:31 PMbland-morning-36590
05/17/2022, 3:55 AMcreamy-smartphone-10810
05/17/2022, 1:14 PMExecuting Step 3/4: DeleteLegacyGraphRelationshipStep...
Failed to delete legacy data from graph: java.lang.ClassCastException: com.linkedin.metadata.graph.elastic.ElasticSearchGraphService cannot be cast to com.linkedin.metadata.graph.neo4j.Neo4jGraphService
Failed to delete legacy data from graph: java.lang.ClassCastException: com.linkedin.metadata.graph.elastic.ElasticSearchGraphService cannot be cast to com.linkedin.metadata.graph.neo4j.Neo4jGraphService
Failed Step 3/4: DeleteLegacyGraphRelationshipStep. Failed after 1 retries.
I’m using elasticsearch
as graph_service_impl, any idea of what could be happening?red-napkin-59945
05/17/2022, 4:13 PMprehistoric-room-17640
05/17/2022, 9:04 PMprehistoric-room-17640
05/17/2022, 9:04 PMrich-policeman-92383
05/18/2022, 7:05 PMgreat-cpu-72376
05/19/2022, 9:59 AMquery {
search(input: {type: DATA_PLATFORM, query: "*"}){
total
searchResults{
entity{
urn
}
}
}
}
what is wrong?
I receive always 0 but there are at list one data platform: PostgreSQL. If I write something in query it is the same. I am very new in datahub and graphqlgreat-cpu-72376
05/19/2022, 3:35 PM[
{
"aspect": {
"corpUserKey":{
"username": "xxx"
},
"corpUserInfo":{
"firstName": "XXX First",
"lastName": "XXXX",
"countryCode": "IT"
},
"corpUserEditableInfo":{
"displayName": "Giorgio",
"aboutMe": "I am trying to add this user",
"teams": ["it-svc-app"],
"skills": ["sql"],
"title": "Data Architect",
"email": "<mailto:giorgio@giorgi.net|giorgio@giorgi.net>"
},
"corpUserStatus":{
"status": "ACTIVE",
}
},
"entityType": "CorpUser",
"entityUrn": "urn:li:CorpUser:xxx"
}
]
There is the aspect dict, with the aspects: corpUserKey, corpUserInfo and corpUserEditableInfo and the entity type and the entityrn. I execute a post to the gms/openapi/entitites/v1 but I receive this error:
15:34:14.631 [qtp544724190-21] WARN o.s.w.s.m.s.DefaultHandlerExceptionResolver:208 - Resolved [org.springframework.http.converter.HttpMessageNotReadableException: JSON parse error: Unexpected character ('}' (code 125)): was expecting double-quote to start field name; nested exception is com.fasterxml.jackson.databind.JsonMappingException: Unexpected character ('}' (code 125)): was expecting double-quote to start field name<EOL> at [Source: (org.springframework.util.StreamUtils$NonClosingInputStream); line: 24, column: 5] (through reference chain: java.util.ArrayList[0]->io.datahubproject.openapi.dto.UpsertAspectRequest$UpsertAspectRequestBuilder["aspect"])]
I validated the json with a parser, what is the problem?mysterious-butcher-86719
05/19/2022, 3:40 PMvictorious-pager-14424
05/19/2022, 6:48 PM./gradlew :datahub-frontend:dist -x yarnTest -x yarnLint
, it fails during the :datahub-web-react:yarnGenerate
step. More info in đź§µmicroscopic-mechanic-13766
05/20/2022, 10:56 AMFailed to create consumer: No provider for SASL mechanism GSSAPI: recompile librdkafka with libsasl2 or openssl support. Current build options: PLAIN SASL_SCRAM OAUTHBEARER"}
gentle-camera-33498
05/20/2022, 3:11 PMgifted-bird-57147
05/21/2022, 9:54 AMdatahub docker nuke --keep-data
datahub docker quickstart
But doing so results in the following error:
ERROR: for datahub-frontend-react Cannot start service datahub-frontend-react: driver failed programming external connectivity on endpoint datahub-frontend-react (e75695d412987a9e3b70806a9905d0798a64bf6d52c1e0afd9cd626c1895a5bf): Error starting userland proxy: listen tcp4 0.0.0.0:9002: bind: address already in use
Any help solving this would be appreciated! (i'm not aware of anything else running on port 9002...)straight-wire-61463
05/23/2022, 12:50 AM