creamy-pizza-80433
06/22/2023, 8:56 AMdazzling-rainbow-96194
06/22/2023, 4:48 PM2023-06-22 16:40:46,537 [I/O dispatcher 1] ERROR c.l.m.s.e.update.BulkListener:56 - Error feeding bulk request. No retries left. Request: Failed to perform bulk request: index [dh_containerindex_v2], optype: [UPDATE], type [_doc], id [urn%3Ali%3Acontainer%3A47eb8e03b73baa9876828b6d3649509c];Failed to perform bulk request: index [dh_containerindex_v2], optype: [UPDATE], type [_doc], id [urn%3Ali%3Acontainer%3A47eb8e03b73baa9876828b6d3649509c];Failed to perform bulk request: index [dh_containerindex_v2], optype: [UPDATE], type [_doc], id [urn%3Ali%3Acontainer%3A47eb8e03b73baa9876828b6d3649509c];Failed to perform bulk request: index [dh_containerindex_v2], optype: [UPDATE], type [_doc], id [urn%3Ali%3Acontainer%3A47eb8e03b73baa9876828b6d3649509c];
anyone has seen this before? Any tips on how to resolve this?numerous-address-22061
06/22/2023, 9:08 PMred-sundown-5665
06/23/2023, 9:40 AMglamorous-spring-97970
06/23/2023, 9:48 AMglamorous-spring-97970
06/23/2023, 9:49 AMglamorous-spring-97970
06/23/2023, 9:49 AMacceptable-computer-51491
06/23/2023, 10:14 AMdatahub-elasticsearch-setup-job
pod is getting 403 response when it tries to connect with open search.
Any idea why is this happening ?quiet-businessperson-49384
06/23/2023, 12:35 PMsalmon-area-51650
06/23/2023, 12:37 PMdatahub-gms
but I’m stuck. So, I currently have the following configuration in my yaml file:
datahub-gms:
enabled: true
image:
repository: linkedin/datahub-gms
tag: "v0.10.3"
service:
type: ClusterIP
env:
- name: JAVA_OPTS
value: "-Dlog4j.rootLogger=ERROR,stdout"
But I’m still seeing INFO
and WARN
logs.
Any advice?
Thanks!powerful-cat-68806
06/24/2023, 4:48 PMapiVersion: v2
appVersion: v0.9.3
description: A Helm chart for LinkedIn DataHub's datahub-gms component
name: datahub-gms
type: application
version: 0.2.165
Also - how can I found, from the namespace, what’s the GMS version?high-twilight-23787
06/24/2023, 7:11 PM[0]: index [system_metadata_service_v1], type [_doc], id [gPRT5nUbjZTmpbFKO3+1Mw==], message [ElasticsearchException[Elasticsearch exception [type=cluster_block_exception, reason=index [system_metadata_service_v1] blocked by: [TOO_MANY_REQUESTS/12/disk usage exceeded flood-stage watermark, index has read-only-allow-delete block];]]]
[2]: index [graph_service_v1], type [_doc], id [sDI2wkN1U9JE9C/8/heWsw==], message [ElasticsearchException[Elasticsearch exception [type=cluster_block_exception, reason=index [graph_service_v1] blocked by: [TOO_MANY_REQUESTS/12/disk usage exceeded flood-stage watermark, index has read-only-allow-delete block];]]]
fierce-night-40574
06/25/2023, 2:35 AMfuture-holiday-32084
06/25/2023, 7:56 AMrich-policeman-92383
06/25/2023, 8:32 AMclever-magician-79463
06/25/2023, 8:36 AMsalmon-area-51650
06/25/2023, 10:16 AMdbt
as I cannot see the dbt test
executions. Evaluations are always empty. For example, this is the content of run_results.json
:
"unique_id": "test.snowflake_db_transformations.equality_caregiver_check_in_and_check_outs_source_ref_caregiver_check_in_and_check_outs_tests_target____tests.ffe5aef1cc"}, {"status": "success", "timing": [{"name": "compile", "started_at": "2023-06-24T12:41:54.699195Z", "completed_at": "2023-06-24T12:41:54.720918Z"}, {"name": "execute", "started_at": "2023-06-24T12:41:54.725765Z", "completed_at": "2023-06-24T12:41:54.725780Z"}], "thread_id": "Thread-4", "execution_time": 0.04546833038330078, "adapter_response": {}, "message": null, "failures": null, "unique_id":
Attached the output of the metadata ingestion job.aloof-energy-17918
06/26/2023, 2:44 AMblue-rainbow-97669
06/26/2023, 7:06 AMmutation createGlossaryNode($name: String!, $id: String!,$parentNode: String!)
{
createGlossaryNode
(
input:
{name: $name, id: $id, parentNode: $parentNode}
)
}
{
"name": "TestingMankamalL1-new3",
"id": "TestingMankamalL1-new3",
"parentNode": "urn:li:glossaryNode:TestingMankamalL1-new2"
}
Error:
{
"errors": [
{
"message": "An unknown error occurred.",
"locations": [
{
"line": 34,
"column": 3
}
],
"path": [
"createGlossaryNode"
],
"extensions": {
"code": 500,
"type": "SERVER_ERROR",
"classification": "DataFetchingException"
}
}
],
"data": {
"createGlossaryNode": null
},
"extensions": {}
}
orange-gpu-90973
06/26/2023, 7:55 AMelegant-salesmen-99143
06/26/2023, 2:00 PMwide-florist-83539
06/26/2023, 10:30 PMERROR - Error sending metadata to datahub: ('Unable to emit metadata to DataHub GMS: Invalid format for aspect: {inputDatajobs=[], inputDatasets=[urn:li:dataset:(urn:li:dataPlatform:snowflake,DEV.ACCOUNTS,dev)
Am I not allowed to add the environment per dataset emit event?
I see that this is still a property in the Entity class for Dataset. Here is my code btw
Dataset(
"snowflake",
str(
snowflake_get_database(env=ENV) + "." + SCHEMA + ".ACCOUNTS"
).upper(),
"DEV"
)
I assume its probably easier to just modify the datahub.cluster
default value to dev so all metadata events emitted from airflow are already labeled Dev.mysterious-wolf-37802
06/27/2023, 3:26 AMrich-restaurant-61261
06/27/2023, 4:51 AM[Errno 8] nodename nor servname provided, or not known
delightful-autumn-14108
06/27/2023, 11:55 AMcuddly-butcher-39945
06/27/2023, 2:44 PMfierce-restaurant-41034
06/27/2023, 4:14 PMdazzling-rainbow-96194
06/27/2023, 8:12 PMrich-restaurant-61261
06/27/2023, 9:16 PMrich-restaurant-61261
06/28/2023, 12:50 AM