happy-easter-36246
03/30/2023, 3:22 AMnumerous-account-62719
03/30/2023, 7:22 AMred-plumber-64268
03/30/2023, 7:59 AMgifted-diamond-19544
03/30/2023, 8:46 AMfresh-cricket-75926
03/30/2023, 11:17 AMfancy-shoe-14428
03/30/2023, 12:43 PMdatahub-actions
container uses so much space? I am using the quickstart image and it rapidly increased to 15gb when I tried to ingest the tables I have on redshift… And I am not even using actions 😆 Any help would be appreciated 🫶blue-honey-61652
03/30/2023, 1:08 PMmost-nightfall-36645
03/30/2023, 2:12 PM2023-03-30 13:40:08.175 ERROR 1 --- [ad | producer-1] c.l.m.dao.producer.KafkaHealthChecker : Failed to emit MCL for entity urn:li:dataset:(urn:li:dataPlatform:XXXXXXX,PROD)
org.apache.kafka.common.errors.RecordTooLargeException: The request included a message larger than the max message size the server will accept.
I set the brokers max.message.bytes
and replica.fetch.max.bytes
to 1GB (much larger than intended final config) and set the emitter jobs SPRING_KAFKA_PRODUCER_PROPERTIES_MAX_REQUEST_SIZE
environment variable to 1GB.
However I still can emit data. I checked the size of the meta data and it is around 1.4MB.
Am I missing something?kind-lifeguard-14131
03/30/2023, 2:18 PMloud-hospital-37195
03/30/2023, 2:30 PMsome-mouse-2910
03/30/2023, 5:41 PM/ingestion
site renders a white background throwing the following error in the console:
Uncaught (in promise) TypeError: Cannot read properties of undefined (reading 'urn')
.
Ingesting datasets stopped working. Do we have to nuke our database and re-i-ngest all of our sources? Do you see any ways to recover from that?
We want to roll out datahub to our 20k employees and I am right now ingesting APIs etc, but there seem to be some difficulties with that.
Thank you for a response.
The full stack trace:
react-dom.production.min.js:216 TypeError: Cannot read properties of undefined (reading 'urn')
at IngestionSourceTable.tsx:116:113
at Array.map (<anonymous>)
at YU (IngestionSourceTable.tsx:107:31)
wide-optician-47025
03/30/2023, 5:51 PMminiature-journalist-76345
03/31/2023, 7:28 AMlimited-library-89060
03/31/2023, 10:15 AMfailing
. Any idea how to show all the test results if all the tests are passed ?white-grass-55842
03/31/2023, 12:54 PMabundant-airport-72599
03/31/2023, 5:56 PMshy-dog-84302
04/01/2023, 3:15 AMio.acryl:datahub-client:0.10.1
) reports security vulnerabilities. Is there any plan/workaround to fix this?
datahub-client-0.10.1.jar/META-INF/maven/org.apache.avro/avro/pom.xml (pkg:maven/org.apache.avro/avro@1.7.7, cpe:2.3:a:apache:avro:1.7.7:*:*:*:*:*:*:*) : CVE-2021-43045
datahub-client-0.10.1.jar/META-INF/maven/org.apache.commons/commons-text/pom.xml (pkg:maven/org.apache.commons/commons-text@1.8, cpe:2.3:a:apache:commons_text:1.8:*:*:*:*:*:*:*) : CVE-2022-42889
future-florist-65080
04/02/2023, 10:48 PMNo Role
.
Is it possible to have all users default to Reader
?few-sunset-43876
04/03/2023, 4:29 AMdocker system prune
and restart Elasticsearch and gms container to clean up.
docker system df:
TYPE TOTAL ACTIVE SIZE RECLAIMABLE
Images 25 8 16.19GB 7.369GB (45%)
Containers 8 8 929.7MB 0B (0%)
Local Volumes 147 9 26.47GB 24.84GB (93%)
Build Cache 25 0 0B 0B
The logs from Elasticsearch:
{"type": "server", "timestamp": "2023-04-03T04:25:30,850Z", "level": "WARN", "component": "o.e.c.r.a.DiskThresholdMonitor", "cluster.name": "docker-cluster", "node.name": "elasticsearch", "message": "flood stage disk watermark [95%] exceeded on [CA9sNvGOSr2sNBFRMst0JQ][elasticsearch][/usr/share/elasticsearch/data/nodes/0] free: 9.3gb[1.9%], all indices on this node will be marked read-only", "cluster.uuid": "M3xdtmw8TFCGL_RqIP650Q", "node.id": "CA9sNvGOSr2sNBFRMst0JQ" }
{"type": "server", "timestamp": "2023-04-03T04:26:00,855Z", "level": "WARN", "component": "o.e.c.r.a.DiskThresholdMonitor", "cluster.name": "docker-cluster", "node.name": "elasticsearch", "message": "flood stage disk watermark [95%] exceeded on [CA9sNvGOSr2sNBFRMst0JQ][elasticsearch][/usr/share/elasticsearch/data/nodes/0] free: 9.3gb[1.9%], all indices on this node will be marked read-only", "cluster.uuid": "M3xdtmw8TFCGL_RqIP650Q", "node.id": "CA9sNvGOSr2sNBFRMst0JQ" }
{"type": "server", "timestamp": "2023-04-03T04:26:30,857Z", "level": "WARN", "component": "o.e.c.r.a.DiskThresholdMonitor", "cluster.name": "docker-cluster", "node.name": "elasticsearch", "message": "flood stage disk watermark [95%] exceeded on [CA9sNvGOSr2sNBFRMst0JQ][elasticsearch][/usr/share/elasticsearch/data/nodes/0] free: 9.3gb[1.9%], all indices on this node will be marked read-only", "cluster.uuid": "M3xdtmw8TFCGL_RqIP650Q", "node.id": "CA9sNvGOSr2sNBFRMst0JQ" }
Logs from gms:
04:26:25.397 [ThreadPoolTaskExecutor-1] INFO c.l.m.k.h.i.IngestionSchedulerHook:56 - Received UPSERT to Ingestion Source. Rescheduling the source (if applicable). urn: urn:li:dataHubIngestionSource:3b33a8fc-b106-460f-b90b-3ca816c77910, key: {value=ByteString(length=45,bytes=7b226964...3130227d), contentType=application/json}.
04:26:25.398 [ThreadPoolTaskExecutor-1] INFO c.d.m.ingestion.IngestionScheduler:105 - Unscheduling ingestion source with urn urn:li:dataHubIngestionSource:3b33a8fc-b106-460f-b90b-3ca816c77910
04:26:25.399 [ThreadPoolTaskExecutor-1] INFO c.d.m.ingestion.IngestionScheduler:138 - Scheduling next execution of Ingestion Source with urn urn:li:dataHubIngestionSource:3b33a8fc-b106-460f-b90b-3ca816c77910. Schedule: 0 1 * * *
04:26:25.400 [ThreadPoolTaskExecutor-1] INFO c.d.m.ingestion.IngestionScheduler:167 - Scheduled next execution of Ingestion Source with urn urn:li:dataHubIngestionSource:3b33a8fc-b106-460f-b90b-3ca816c77910 in 48814600ms.
04:26:27.448 [ThreadPoolTaskExecutor-1] INFO c.l.m.k.t.DataHubUsageEventTransformer:74 - Invalid event type: CreateIngestionSourceEvent
04:26:27.449 [ThreadPoolTaskExecutor-1] WARN c.l.m.k.DataHubUsageEventsProcessor:56 - Failed to apply usage events transform to record: {"type":"CreateIngestionSourceEvent","sourceType":"bigquery","interval":"0 1 * * *","actorUrn":"urn:li:corpuser:datahub","timestamp":1680495987418,"date":"Mon Apr 03 2023 11:26:27 GMT+0700 (Indochina Time)","userAgent":"Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/107.0.0.0 Safari/537.36","browserId":"c7e9ab30-25a7-4874-a4e6-f2c694501ccc"}
04:26:27.474 [qtp522764626-22] INFO c.l.m.r.entity.AspectResource:143 - INGEST PROPOSAL proposal: {aspectName=dataHubExecutionRequestResult, entityKeyAspect={contentType=application/json, value=ByteString(length=46,bytes=7b226964...3463227d)}, entityType=dataHubExecutionRequest, aspect={contentType=application/json, value=ByteString(length=51,bytes=7b227374...3437307d)}, changeType=UPSERT}
04:26:27.485 [pool-12-thread-1] INFO c.l.m.filter.RestliLoggingFilter:55 - POST /aspects?action=ingestProposal - ingestProposal - 200 - 11ms
04:26:27.495 [ThreadPoolTaskExecutor-1] INFO c.l.m.k.t.DataHubUsageEventTransformer:74 - Invalid event type: ExecuteIngestionSourceEvent
04:26:27.495 [ThreadPoolTaskExecutor-1] WARN c.l.m.k.DataHubUsageEventsProcessor:56 - Failed to apply usage events transform to record: {"type":"ExecuteIngestionSourceEvent","actorUrn":"urn:li:corpuser:datahub","timestamp":1680495987487,"date":"Mon Apr 03 2023 11:26:27 GMT+0700 (Indochina Time)","userAgent":"Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/107.0.0.0 Safari/537.36","browserId":"c7e9ab30-25a7-4874-a4e6-f2c694501ccc"}
04:26:28.968 [I/O dispatcher 1] ERROR c.l.m.s.e.update.BulkListener:25 - Failed to feed bulk request. Number of events: 4 Took time ms: -1 Message: failure in bulk execution:
[0]: index [datahubingestionsourceindex_v2], type [_doc], id [urn%3Ali%3AdataHubIngestionSource%3A3b33a8fc-b106-460f-b90b-3ca816c77910], message [ElasticsearchException[Elasticsearch exception [type=cluster_block_exception, reason=index [datahubingestionsourceindex_v2] blocked by: [TOO_MANY_REQUESTS/12/disk usage exceeded flood-stage watermark, index has read-only-allow-delete block];]]]
[1]: index [system_metadata_service_v1], type [_doc], id [D9J9LxuD6yBN4lJYY0FaMg==], message [ElasticsearchException[Elasticsearch exception [type=cluster_block_exception, reason=index [system_metadata_service_v1] blocked by: [TOO_MANY_REQUESTS/12/disk usage exceeded flood-stage watermark, index has read-only-allow-delete block];]]]
[2]: index [datahubingestionsourceindex_v2], type [_doc], id [urn%3Ali%3AdataHubIngestionSource%3A3b33a8fc-b106-460f-b90b-3ca816c77910], message [ElasticsearchException[Elasticsearch exception [type=cluster_block_exception, reason=index [datahubingestionsourceindex_v2] blocked by: [TOO_MANY_REQUESTS/12/disk usage exceeded flood-stage watermark, index has read-only-allow-delete block];]]]
[3]: index [system_metadata_service_v1], type [_doc], id [F7B9+ecZWomYYe5LYipdcA==], message [ElasticsearchException[Elasticsearch exception [type=cluster_block_exception, reason=index [system_metadata_service_v1] blocked by: [TOO_MANY_REQUESTS/12/disk usage exceeded flood-stage watermark, index has read-only-allow-delete block];]]]
04:26:29.528 [qtp522764626-17] INFO c.l.m.r.entity.AspectResource:143 - INGEST PROPOSAL proposal: {aspectName=dataHubExecutionRequestResult, entityKeyAspect={contentType=application/json, value=ByteString(length=46,bytes=7b226964...3463227d)}, entityType=dataHubExecutionRequest, aspect={contentType=application/json, value=ByteString(length=389,bytes=7b227374...3437307d)}, changeType=UPSERT}
04:26:29.555 [pool-12-thread-1] INFO c.l.m.filter.RestliLoggingFilter:55 - POST /aspects?action=ingestProposal - ingestProposal - 200 - 27ms
04:26:29.749 [qtp522764626-17] INFO c.l.m.r.entity.AspectResource:93 - GET ASPECT urn: urn:li:telemetry:clientId aspect: telemetryClientId version: 0
04:26:29.752 [pool-12-thread-1] INFO c.l.m.filter.RestliLoggingFilter:55 - GET /aspects/urn%3Ali%3Atelemetry%3AclientId?aspect=telemetryClientId&version=0 - get - 200 - 3ms
04:26:31.566 [qtp522764626-23] INFO c.l.m.r.entity.AspectResource:143 - INGEST PROPOSAL proposal: {aspectName=dataHubExecutionRequestResult, entityKeyAspect={contentType=application/json, value=ByteString(length=46,bytes=7b226964...3463227d)}, entityType=dataHubExecutionRequest, aspect={contentType=application/json, value=ByteString(length=654,bytes=7b227374...3437307d)}, changeType=UPSERT}
04:26:31.986 [I/O dispatcher 1] ERROR c.l.m.s.e.update.BulkListener:25 - Failed to feed bulk request. Number of events: 8 Took time ms: -1 Message: failure in bulk execution:
[0]: index [datahubexecutionrequestindex_v2], type [_doc], id [urn%3Ali%3AdataHubExecutionRequest%3A67e66178-9cf9-40ac-8496-ac26761ba24c], message [ElasticsearchException[Elasticsearch exception [type=cluster_block_exception, reason=index [datahubexecutionrequestindex_v2] blocked by: [TOO_MANY_REQUESTS/12/disk usage exceeded flood-stage watermark, index has read-only-allow-delete block];]]]
[1]: index [datahubexecutionrequestindex_v2], type [_doc], id [urn%3Ali%3AdataHubExecutionRequest%3A67e66178-9cf9-40ac-8496-ac26761ba24c], message [ElasticsearchException[Elasticsearch exception [type=cluster_block_exception, reason=index [datahubexecutionrequestindex_v2] blocked by: [TOO_MANY_REQUESTS/12/disk usage exceeded flood-stage watermark, index has read-only-allow-delete block];]]]
[2]: index [graph_service_v1], type [_doc], id [t/DPfLYDIXzDBvDNKYzsBA==], message [ElasticsearchException[Elasticsearch exception [type=cluster_block_exception, reason=index [graph_service_v1] blocked by: [TOO_MANY_REQUESTS/12/disk usage exceeded flood-stage watermark, index has read-only-allow-delete block];]]]
[3]: index [system_metadata_service_v1], type [_doc], id [mK6kjzTpk8iRRsx5Eg7tJA==], message [ElasticsearchException[Elasticsearch exception [type=cluster_block_exception, reason=index [system_metadata_service_v1] blocked by: [TOO_MANY_REQUESTS/12/disk usage exceeded flood-stage watermark, index has read-only-allow-delete block];]]]
[4]: index [system_metadata_service_v1], type [_doc], id [yss7L2uhB/qEk71o8Vkc8w==], message [ElasticsearchException[Elasticsearch exception [type=cluster_block_exception, reason=index [system_metadata_service_v1] blocked by: [TOO_MANY_REQUESTS/12/disk usage exceeded flood-stage watermark, index has read-only-allow-delete block];]]]
[5]: index [datahubexecutionrequestindex_v2], type [_doc], id [urn%3Ali%3AdataHubExecutionRequest%3A67e66178-9cf9-40ac-8496-ac26761ba24c], message [ElasticsearchException[Elasticsearch exception [type=cluster_block_exception, reason=index [datahubexecutionrequestindex_v2] blocked by: [TOO_MANY_REQUESTS/12/disk usage exceeded flood-stage watermark, index has read-only-allow-delete block];]]]
[6]: index [datahubexecutionrequestindex_v2], type [_doc], id [urn%3Ali%3AdataHubExecutionRequest%3A67e66178-9cf9-40ac-8496-ac26761ba24c], message [ElasticsearchException[Elasticsearch exception [type=cluster_block_exception, reason=index [datahubexecutionrequestindex_v2] blocked by: [TOO_MANY_REQUESTS/12/disk usage exceeded flood-stage watermark, index has read-only-allow-delete block];]]]
[7]: index [system_metadata_service_v1], type [_doc], id [QVHUM/x8ssrKDaAcXWOgng==], message [ElasticsearchException[Elasticsearch exception [type=cluster_block_exception, reason=index [system_metadata_service_v1] blocked by: [TOO_MANY_REQUESTS/12/disk usage exceeded flood-stage watermark, index has read-only-allow-delete block];]]]
04:26:31.987 [pool-12-thread-1] INFO c.l.m.filter.RestliLoggingFilter:55 - POST /aspects?action=ingestProposal - ingestProposal - 200 - 421ms
I run datahub locally. Is it because of resource issue? how can I check and solve it?
Thanks in advance!victorious-planet-2053
04/03/2023, 10:12 AMdatahub docker quickstart
and recive error:
Unable to run quickstart - the following issues were detected:
- datahub-frontend-react is running by not yet healthy
- datahub-gms is still starting
- broker is not running
- elasticsearch-setup is still running
- elasticsearch is running by not yet healthy
Can someone help me please?helpful-quill-60747
04/03/2023, 10:47 AMstale-minister-18858
04/03/2023, 12:39 PMagreeable-belgium-70840
04/03/2023, 3:02 PMclient.dns.lookup = use_all_dns_ips
That used to be
client.dns.lookup = default
Could it be the case? How can I change that? I tried passing the env variable below, but there wasn't any change:
- name: KAFKA_CLIENT_DNS_LOOKUP
value: "default"
Moreover, the old one used to be like this:
ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
How can I add TLSv1 support to the new one?
Regards,
Yiannicalm-balloon-31412
04/03/2023, 8:37 PMANTLR Tool version 4.5 used for code generation does not match the current runtime version 4.8ANTLR Runtime version 4.5 used for parser compilation does not match the current runtime version 4.8ANTLR Tool version 4.5 used for code generation does not match the current runtime version 4.8ANTLR Runtime version 4.5 used for parser compilation does not match the current runtime version 4.82023/03/20 18:54:37 Command exited with error: exit status 1
when I am trying to replace datahub-gms container with local code:
(cd docker && COMPOSE_DOCKER_CLI_BUILD=1 DOCKER_BUILDKIT=1 docker-compose -p datahub -f docker-compose-without-neo4j.yml -f docker-compose-without-neo4j.override.yml -f docker-compose.dev.yml up -d --no-deps --force-recreate datahub-gms)
Any recommendations? I've tried looking through other threads but have not found anything usefulagreeable-table-54007
04/04/2023, 9:14 AMdatahub docker quickstart
(windows VM and datahub version 0.10.1) and receive errors:
Unable to run quickstart - the following issues were detected:
- datahub-upgrade is still running
- datahub-gms is still starting
- elasticsearch-setup is still running
- mysql-setup is still running
- elasticsearch is running by not yet healthy
4months ago to discover datahub, it worked and now it's not so i did a nuke / prune and did the quickstart guide again but got the previous errors..
Can someone help me please?quick-pizza-8906
04/04/2023, 2:26 PMwonderful-quill-11255
04/04/2023, 5:45 PMclever-spring-20900
04/04/2023, 11:05 PMastonishing-dusk-99990
04/05/2023, 2:11 AMhelm upgrade chart
my two pods error which are prerequisites-kafka and kafka-setup-job.
I attached screenshoot and logs for two pods in case everyone know how to solve it
Thank you.
Notes:
• Image datahub v0.10.0busy-analyst-35820
04/05/2023, 7:02 AM