better-fireman-33387
05/03/2023, 6:25 AMminiature-lighter-59048
05/03/2023, 7:33 AMwide-ghost-47822
05/03/2023, 7:33 AMwide-afternoon-79955
05/03/2023, 12:52 PMOpenSearch 1.3
) and we are facing 500s while downloading a lineage data for a Swnoflake table. Error stack trace is in thread.magnificent-lawyer-97772
05/03/2023, 1:47 PMbest-umbrella-88325
05/03/2023, 3:25 PMelegant-salesmen-99143
05/03/2023, 4:14 PMlively-dusk-19162
05/03/2023, 5:22 PMbland-orange-13353
05/03/2023, 6:46 PMrich-state-73859
05/03/2023, 6:55 PMripe-oxygen-26489
05/03/2023, 7:05 PMfast-advantage-32018
05/03/2023, 9:07 PM'str' object is not callable
error. I thought I had downgraded the Datahub CLI version to .10.1.2
version, but am still running into the error. Is there something I need to change in the docker-compose.yml file? Any help would be appreciated.creamy-ram-28134
05/03/2023, 9:22 PMcreamy-ram-28134
05/03/2023, 9:22 PM[root@adkube06 ~]# kubectl logs -f datahub-datahub-gms-b4c458457-qxzpb -n gopikab
2023/05/03 21:19:19 Waiting for: <tcp://prerequisites-mysql:3306>
2023/05/03 21:19:19 Waiting for: <tcp://prerequisites-kafka:9092>
2023/05/03 21:19:19 Waiting for: <http://elasticsearch-master:9200>
2023/05/03 21:19:19 Waiting for: http:
2023/05/03 21:19:19 Problem with request: Get http:: http: no Host in request URL. Sleeping 1s
2023/05/03 21:19:19 Connected to <tcp://prerequisites-mysql:3306>
2023/05/03 21:19:19 Connected to <tcp://prerequisites-kafka:9092>
2023/05/03 21:19:19 Received 200 from <http://elasticsearch-master:9200>
2023/05/03 21:19:20 Problem with request: Get http:: http: no Host in request URL. Sleeping 1s
2023/05/03 21:19:21 Problem with request: Get http:: http: no Host in request URL. Sleeping 1s
2023/05/03 21:19:22 Problem with request: Get http:: http: no Host in request URL. Sleeping 1s
2023/05/03 21:19:23 Problem with request: Get http:: http: no Host in request URL. Sleeping 1s
creamy-ram-28134
05/03/2023, 9:23 PMcreamy-ram-28134
05/03/2023, 9:24 PMpod/prerequisites-cp-schema-registry-6c69f9f665-mbq9k 1/2 CrashLoopBackOff 25 (3m7s ago) 106m
[root@adkube06 ~]# kubectl logs -f pod/prerequisites-cp-schema-registry-6c69f9f665-mbq9k -n gopikab
Defaulted container "prometheus-jmx-exporter" out of: prometheus-jmx-exporter, cp-schema-registry-server
Unrecognized VM option 'UseCGroupMemoryLimitForHeap'
Error: Could not create the Java Virtual Machine.
Error: A fatal exception has occurred. Program will exit.
few-sunset-43876
05/04/2023, 4:18 AM"Cannot handle <project-id>.<dataset>.<table-id>$__PARTITIONS_SUMMARY__ - poorly formatted table name, contains ['$']"
I see the following commit has fixed it. But it still happens to me. Can anyone help? Thanks!
https://github.com/datahub-project/datahub/pull/3842adorable-megabyte-63781
05/04/2023, 6:10 AMCould not resolve all artifacts for configuration ':classpath'.> Could not resolve com.linkedin.pegasusgradle plugins29.22.16. Required by: project : > Could not resolve com.linkedin.pegasusgradle plugins29.22.16. > Could not get resource 'https://packages.confluent.io/maven/com/linkedin/pegasus/gradle-plugins/29.22.16/gradle-plugins-29.22.16.pom'. > Could not GET 'https://packages.confluent.io/maven/com/linkedin/pegasus/gradle-plugins/29.22.16/gradle-plugins-29.22.16.pom'. > Connection reset > Could not resolve com.linkedin.pegasusgradle plugins29.22.16. > Could not get resource 'https://linkedin.jfrog.io/artifactory/open-source/com/linkedin/pegasus/gradle-plugins/29.22.16/gradle-plugins-29.22.16.pom'. > Could not GET 'https://linkedin.jfrog.io/artifactory/open-source/com/linkedin/pegasus/gradle-plugins/29.22.16/gradle-plugins-29.22.16.pom'. > Connection reset
nice-helmet-40615
05/04/2023, 11:42 AMmany-glass-1784
05/04/2023, 3:03 PMCaused by: java.lang.RuntimeException: Failed to batch load data products
at com.linkedin.datahub.graphql.types.dataproduct.DataProductType.batchLoad(DataProductType.java:81)
at com.linkedin.datahub.graphql.GmsGraphQLEngine.lambda$createDataLoader$194(GmsGraphQLEngine.java:1684)
... 2 common frames omitted
Caused by: java.lang.IllegalArgumentException: Failed to convert urn to entity key: urns parts and key fields do not have same length for urn:li:dataProduct:be.publiq.vrijetijdsparticipatie-publiq-uit-locaties
at com.linkedin.metadata.utils.EntityKeyUtils.convertUrnToEntityKey(EntityKeyUtils.java:95)
at com.linkedin.metadata.entity.EntityService.getKeyEnvelopedAspect(EntityService.java:1891)
at com.linkedin.metadata.entity.EntityService.getCorrespondingAspects(EntityService.java:393)
at com.linkedin.metadata.entity.EntityService.getLatestEnvelopedAspects(EntityService.java:336)
at com.linkedin.metadata.entity.EntityService.getEntitiesV2(EntityService.java:292)
at com.linkedin.metadata.client.JavaEntityClient.batchGetV2(JavaEntityClient.java:111)
at com.linkedin.datahub.graphql.types.dataproduct.DataProductType.batchLoad(DataProductType.java:63)
... 3 common frames omitted
After some looking around I found that there was some data in a hidden ElasticSearch index, .ds-datahub_usage_event-000001
, that still seemed to contain rows with URNs that refer to data that was deleted. it looks like the recommendations/suggestions section on the main page uses this index and as such gets some data that contains entity URNs in an old and invalid format which causes exceptions.
When I use the ElasticSearch APIs to clear this index, the error goes away, and when clicking around in the application it gets filled again, this time with data with valid URNs.
Is it normal that this index isn't kept up to data with the normal entity deletes (as the database and non-hidden ElasticSearch indices are actually cleared)? If it is normal, is there a better way to clear the indices? Locally the clear of the hidden index works, but on our actual deployed environment this only clears some of the errors, but even with the deletes and the cleared index it still seems to find some invalid data somewhere?big-ocean-9800
05/04/2023, 6:15 PMv0.8.41
to v0.9.5
.
The main issue we are seeing is that browse paths are not working as expected.
Before, the browse path was properly being populated based on instance name and database name, now it’s being truncated to just the platform name after we run an ingest.
Here’s an example of two different versions of the browsePaths
aspect for the same URN before and after the upgrade:
```
-[ RECORD 1 ]--+--------------------------------------------------------------------------------------------------------------------------------------------------------
urn | urn:li:dataset:(urn:li:dataPlatform:CockroachDB,test-instance.test-database.test-schema.test-table,PROD)
aspect | browsePaths
version | 0
metadata | {"paths":["/prod/cockroachdb"]}
systemmetadata | {}
createdon | 2023-01-05 17:35:19.596
createdby | urn:li:corpuser:__datahub_system
createdfor |
-[ RECORD 2 ]--+--------------------------------------------------------------------------------------------------------------------------------------------------------
urn | urn:li:dataset:(urn:li:dataPlatform:CockroachDB,test-instance.test-database.test-schema.test-table,PROD)
aspect | browsePaths
version | 1
metadata | {"paths":["/prod/cockroachdb/test-instance/test-database/test-schema/test-table"]}
systemmetadata | {}
createdon | 2022-07-29 22:08:16.392
createdby | urn:li:corpuser:__datahub_system
createdfor |
```
Here you can see that the browse path is no longer being populated the same way, even though the ingestion source code has not changed.
Does anyone know if we are missing something around this browsePaths issue? My understanding is that nothing should have changed around this after the version upgrade. I’ve been looking through the ingestion source code but haven’t found any smoking guns yet.creamy-ram-28134
05/04/2023, 7:04 PM2023-05-04 17:43:27,138 [pool-19-thread-1] ERROR c.l.d.g.a.service.AnalyticsService:264 - Search query failed: Elasticsearch exception [type=index_not_found_exception, reason=no such index [datahub_usage_event]]
My question is who or what needs to create those indexes? Is it one of the automatic jobs?lively-dusk-19162
05/04/2023, 7:47 PMearly-hydrogen-27542
05/04/2023, 8:15 PMdbt platform▾
under search▾
pulling no dbt entities▾
bland-orange-13353
05/04/2023, 11:23 PMimportant-intern-48298
05/05/2023, 2:25 AM./gradlew build
Getting the following error:
FAILURE: Build failed with an exception.
* What went wrong:
Execution failed for task 'datahub web reactyarnGenerate'.
Process 'command '/home/Documents/Repositories/datahub/datahub-web-react/.gradle/yarn/yarn-v1.22.0/bin/yarn'' finished with non-zero exit value 1So far I have tried clearing the yarn cache but still getting the same error. I would really appreciate it if someone can assist. Thanks!
bland-orange-13353
05/05/2023, 2:27 AMadamant-furniture-37835
05/05/2023, 7:31 AMblue-microphone-24514
05/05/2023, 7:42 AMmetadata_service_authentication
, all auth (user/pwd, SSO) stop working with a Provided credentials do not match known system client id & client secret
in the logs ?rapid-forest-41223
05/05/2023, 8:27 AMsearchAcrossEntities
does work and the "enableAutocomplete": true
is set for the field. I do not see the logline Autocompleting query entityName
for the new entity though in the gms.debug.log
after the autoCompleteForMultiple
query is fired.