hallowed-shampoo-52722
12/20/2022, 4:40 PMbland-orange-13353
12/20/2022, 4:40 PMnutritious-bird-77396
12/20/2022, 9:53 PMv0.9.3
version of datahub-gms and v0.9.3
datahub-frontend and 0.0.8
version of datahub actions.
I see the below message when i hit the datahub frontend endpoint - Failed to perform post authentication steps. Error message: com.linkedin.r2.RemoteInvocationException: Failed to get response from server for URI <http://datahub-gms-service>.<team>.svc.cluster.local:8080/aspects
When hitting gms health endpoint returns 200.
Also, datahub frontend /admin endpoint returns GOOD
average-dinner-25106
12/21/2022, 2:32 AMfull-kite-21373
12/21/2022, 9:54 AMfull-kite-21373
12/21/2022, 9:55 AMsquare-solstice-69079
12/21/2022, 11:39 AMdatahub docker quickstart
and
docker-compose -p datahub -f docker-compose.yml -f docker-compose.yml up -d datahub-frontend-react
square-solstice-69079
12/21/2022, 11:41 AMgentle-portugal-21014
12/21/2022, 11:54 AMdamp-greece-27806
12/21/2022, 5:46 PMcode_verifier
parameter that we’re not passing. We’re not seeing how we can pass that. One thing to note/check is that datahub is using pac4j 4.5.7 (https://github.com/datahub-project/datahub/blob/master/build.gradle#L156) and pac4j recently released 5.1.1 with notes about fixing PKCE OIDC flow issues (https://www.pac4j.org/docs/release-notes.html), in case this problem is from an underlying source. Any help you all can provide would be appreciated, our datahub cluster has been down for 2 weeks now because of thisnutritious-bird-77396
12/21/2022, 6:26 PM/login
using admin creds..
But, now its throwing Caused by: java.lang.RuntimeException: Failed to generate session token for user
error
Error stack in 🧵bland-balloon-48379
12/21/2022, 7:49 PMjava.lang.OutOfMemoryError: Java heap space
. When this happens the entire application freezes until those containers are manually restarted. I've tried tripling the RAM for these containers and saw no change.
• When I create a new policy it says it was successfully created, but the new policy does not appear in the list nor do its permissions take effect. When I check the MySQL database though, I can see the rows for the new policy which was created.
I'm unsure if there are other issues, this just what we've observed over the past few days. I downgraded one of our instances to v0.9.2 and all of these issues went away. Are there any thoughts on what may have changed between v0.9.2 and v0.9.3 when could have caused these issues? Also, I saw that v0.9.4 just recently came out. Is there reason to believe that some of these issues may be resolved by upgrading to that version?bitter-waitress-17567
12/22/2022, 5:44 AMbitter-waitress-17567
12/22/2022, 5:44 AM> Task :datahub-web-react:yarnBuild FAILED
FAILURE: Build failed with an exception.
* What went wrong:
Execution failed for task ':datahub-web-react:yarnBuild'.
> Process 'command 'yarn'' finished with non-zero exit value 1
bitter-waitress-17567
12/22/2022, 5:44 AMmicroscopic-mechanic-13766
12/22/2022, 8:48 AMthankful-fireman-70616
12/22/2022, 11:12 AMsteep-midnight-37232
12/22/2022, 1:57 PMbrainy-piano-85560
12/22/2022, 3:09 PM'/usr/local/bin/run_ingest.sh: line 40: 467 Killed ( datahub ${debug_option} ingest run -c "${recipe_file}" '
'${report_option} )\n',
"2022-12-22 14:43:24.766921 [exec_id=b0e5bc27-9d53-46e6-b866-f388b73cac96] INFO: Failed to execute 'datahub ingest'",
'2022-12-22 14:43:24.771317 [exec_id=b0e5bc27-9d53-46e6-b866-f388b73cac96] INFO: Caught exception EXECUTING '
'task_id=b0e5bc27-9d53-46e6-b866-f388b73cac96, name=RUN_INGEST, stacktrace=Traceback (most recent call last):\n'
' File "/usr/local/lib/python3.10/site-packages/acryl/executor/execution/default_executor.py", line 123, in execute_task\n'
' task_event_loop.run_until_complete(task_future)\n'
' File "/usr/local/lib/python3.10/asyncio/base_events.py", line 646, in run_until_complete\n'
' return future.result()\n'
' File "/usr/local/lib/python3.10/site-packages/acryl/executor/execution/sub_process_ingestion_task.py", line 227, in execute\n'
' raise TaskError("Failed to execute \'datahub ingest\'")\n'
"acryl.executor.execution.task.TaskError: Failed to execute 'datahub ingest'\n"]}
Execution finished with errors.
The metadata ingesting worked well. All tables and columns are ingested. Some of the profiling works aswell (the tables that the ingestion finished profiling shows row count etc).
Any thoughts of how I can understand what the problem was?
Thank you.rough-flag-51828
12/23/2022, 9:07 AMRUN_INGEST - {'errors': [],
'exec_id': '5aa6b080-4493-450a-ad72-dc7a7ab9e9ee',
'infos': ['2022-12-23 09:06:34.833273 [exec_id=5aa6b080-4493-450a-ad72-dc7a7ab9e9ee] INFO: Starting execution for task with name=RUN_INGEST',
'2022-12-23 09:06:34.834367 [exec_id=5aa6b080-4493-450a-ad72-dc7a7ab9e9ee] INFO: Caught exception EXECUTING '
'task_id=5aa6b080-4493-450a-ad72-dc7a7ab9e9ee, name=RUN_INGEST, stacktrace=Traceback (most recent call last):\n'
' File "/usr/local/lib/python3.10/site-packages/acryl/executor/execution/default_executor.py", line 123, in execute_task\n'
' task_event_loop.run_until_complete(task_future)\n'
' File "/usr/local/lib/python3.10/asyncio/base_events.py", line 646, in run_until_complete\n'
' return future.result()\n'
' File "/usr/local/lib/python3.10/site-packages/acryl/executor/execution/sub_process_ingestion_task.py", line 73, in execute\n'
' SubProcessTaskUtil._write_recipe_to_file(exec_out_dir, file_name, recipe)\n'
' File "/usr/local/lib/python3.10/site-packages/acryl/executor/execution/sub_process_task_common.py", line 105, in '
'_write_recipe_to_file\n'
' os.makedirs(dir_path, mode = 0o777, exist_ok = True)\n'
' File "/usr/local/lib/python3.10/os.py", line 215, in makedirs\n'
' makedirs(head, exist_ok=exist_ok)\n'
' File "/usr/local/lib/python3.10/os.py", line 225, in makedirs\n'
' mkdir(name, mode)\n'
"PermissionError: [Errno 13] Permission denied: '/tmp/datahub/ingest'\n"]}
Execution finished with errors.
witty-butcher-82399
12/23/2022, 10:29 AMTopicName
and TopicRecordName
strategy are covered (the second in a sort of best-effort mode).
However, when testing, we have found that we miss many schemas that follow the TopicRecordName
strategy.
We had a look to the code here https://github.com/datahub-project/datahub/blob/master/metadata-ingestion/src/datahub/ingestion/source/confluent_schema_registry.py#L72-L78
# Subject name format when the schema registry subject name strategy is
# (a) TopicNameStrategy(default strategy): <topic name>-<key/value>
# (b) TopicRecordNameStrategy: <topic name>-<fully-qualified record name>-<key/value>
for subject in self.known_schema_registry_subjects:
if subject.startswith(topic) and subject.endswith(subject_key_suffix):
return subject
return None
and we think there is a mistake. The code assumes the -(key|value)
suffix also exists for TopicRecordName
strategy, however this is not true; we have plenty of topics with the following format <topic name>-<fully-qualified record name>
for the TopicRecordName
strategy.
Is there any serializer supporting this <topic name>-<fully-qualified record name>-<key/value>
format?
For example, Confluent AVRO serializer doesn’t https://docs.confluent.io/platform/6.1.2/clients/confluent-kafka-python/html/_modules/confluent_kafka/schema_registry/avro.html
As a mitigation, schema resolution may first try exact match for TopicNameStrategy
and if not, do a search only with startswith
(no endswith
). But I think that may bring many false positives and no way to differentiate key and value schemas.
So from my understanding, fetching schemas only works for TopicName
strategy. I’m raising the concern here in case I’m missing something or if someone else has an idea to fix TopicRecordName
strategy. Thanks!
🧵fresh-rocket-98009
12/23/2022, 11:18 AMfaint-actor-78390
12/23/2022, 1:03 PMhandsome-football-66174
12/23/2022, 4:02 PMERROR: Cannot install acryl-datahub[s3]==0.9.4 because these package versions have conflicting dependencies.
The conflict is caused by:
acryl-datahub[s3] 0.9.4 depends on pyspark==3.0.3; extra == "s3"
The user requested (constraint) pyspark==3.2.1
To fix this you could try to:
1. loosen the range of package versions you've specified
2. remove package versions to allow pip attempt to solve the dependency conflict
damp-greece-27806
12/23/2022, 6:03 PM18:01:33 [pool-7-thread-1] ERROR c.d.authorization.DataHubAuthorizer - Failed to retrieve policy urns! Skipping updating policy cache until next refresh. start: 0, count: 30
com.datahub.util.exception.ESQueryException: Search query failed:
at com.linkedin.metadata.search.elasticsearch.query.ESSearchDAO.executeAndExtract(ESSearchDAO.java:73)
at com.linkedin.metadata.search.elasticsearch.query.ESSearchDAO.search(ESSearchDAO.java:100)
at com.linkedin.metadata.search.elasticsearch.ElasticSearchService.search(ElasticSearchService.java:97)
at com.linkedin.metadata.client.JavaEntityClient.search(JavaEntityClient.java:300)
at com.datahub.authorization.PolicyFetcher.fetchPolicies(PolicyFetcher.java:50)
at com.datahub.authorization.PolicyFetcher.fetchPolicies(PolicyFetcher.java:42)
at com.datahub.authorization.DataHubAuthorizer$PolicyRefreshRunnable.run(DataHubAuthorizer.java:223)
at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)
at java.base/java.util.concurrent.FutureTask.runAndReset(FutureTask.java:305)
at java.base/java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:305)
at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
at java.base/java.lang.Thread.run(Thread.java:829)
Caused by: java.lang.RuntimeException: Request cannot be executed; I/O reactor status: STOPPED
at org.elasticsearch.client.RestClient.extractAndWrapCause(RestClient.java:857)
at org.elasticsearch.client.RestClient.performRequest(RestClient.java:259)
at org.elasticsearch.client.RestClient.performRequest(RestClient.java:246)
at org.elasticsearch.client.RestHighLevelClient.internalPerformRequest(RestHighLevelClient.java:1613)
at org.elasticsearch.client.RestHighLevelClient.performRequest(RestHighLevelClient.java:1583)
at org.elasticsearch.client.RestHighLevelClient.performRequestAndParseEntity(RestHighLevelClient.java:1553)
at org.elasticsearch.client.RestHighLevelClient.search(RestHighLevelClient.java:1069)
at com.linkedin.metadata.search.elasticsearch.query.ESSearchDAO.executeAndExtract(ESSearchDAO.java:60)
... 12 common frames omitted
Caused by: java.lang.IllegalStateException: Request cannot be executed; I/O reactor status: STOPPED
at org.apache.http.util.Asserts.check(Asserts.java:46)
at org.apache.http.impl.nio.client.CloseableHttpAsyncClientBase.ensureRunning(CloseableHttpAsyncClientBase.java:90)
at org.apache.http.impl.nio.client.InternalHttpAsyncClient.execute(InternalHttpAsyncClient.java:123)
at org.elasticsearch.client.RestClient.performRequest(RestClient.java:255)
... 18 common frames omitted
bitter-waitress-17567
12/26/2022, 7:16 AMbitter-waitress-17567
12/26/2022, 7:16 AMValidation error (FieldUndefined@[appConfig/viewsConfig]) : Field 'viewsConfig' in type 'AppConfig' is undefined (code undefined)
bitter-waitress-17567
12/26/2022, 7:17 AMbitter-waitress-17567
12/26/2022, 7:17 AMconst config: any = {
amplitude: {
apiKey: 'xxxxxxxxxxxxxxxxxxxx',
},
};
export default config;
bitter-waitress-17567
12/26/2022, 7:17 AM