blue-beach-27940
06/28/2022, 8:05 AMblue-beach-27940
06/28/2022, 8:09 AMblue-beach-27940
06/28/2022, 8:13 AMdatahub docker quickstart
, but got an error blow, anyone has this problem?brainy-crayon-53549
06/28/2022, 1:13 PMmicroscopic-helicopter-87069
06/28/2022, 1:31 PMgreen-lion-58215
06/28/2022, 10:21 PMfrom datahub.ingestion.run.pipeline import Pipeline
pipeline = Pipeline.create(
{
"source": {
"type": "dbt",
"config": {
"manifest_path": "/tmp/manifest.json",
"catalog_path": "/tmp/catalog.json",
"sources_path": "/tmp/sources.json",
"target_platform": "databricks",
"load_schemas": True
},
},
"sink": {
"type": "datahub-rest",
"config": {"server": "http://<masked>:8080"},
},
}
)
Thread in Slack Conversationmany-glass-61317
06/29/2022, 1:22 AMbrainy-crayon-53549
06/29/2022, 11:20 AMtall-fall-45442
06/29/2022, 5:13 PMmysterious-eye-58423
06/29/2022, 9:13 PMThis makes restoring timeseries aspects in a disaster scenario a bit more challenge.
Have we discussed/implemented a solution to recover search index when the attributes of metadata are not persisted in the relational store?lemon-zoo-63387
06/30/2022, 4:11 AMbitter-oxygen-31974
06/30/2022, 4:33 AMbitter-toddler-42943
06/30/2022, 5:54 AMbrainy-crayon-53549
06/30/2022, 11:27 AMmammoth-honey-57770
06/30/2022, 12:03 PMbillions-twilight-48559
06/30/2022, 12:06 PMquick-megabyte-61846
06/30/2022, 12:26 PMcase
Periodic tasks run on airflow with dbt
I was thinking it’s possible to ingest only specific artefacts in my example I would like only to ingest run_results.json
hence test/assertion data to datahub.
With this case, I don’t ingest data that is already in datahub only the assertions which are neededelegant-salesmen-99143
06/30/2022, 1:23 PMcolossal-easter-99672
06/30/2022, 1:29 PMinsert into a select * from k;
insert into b select * from l;
insert into c select * from m;
insert into d select * from n;
insert into e select * from o;
insert into f select * from p;
insert into g select * from q;
insert into h select * from r;
insert into e select * from s;
insert into j select * from t;
if i set to this DataJob outputs = a,b,c,d,e,f,g,h,e,j and inputs = k,l,m,n,o,p,q,r,s,t i get mixed (and wrong) lineage for datasets.
Now i generate fake 10 DataJobs (outputs = a/inputs = k; outputs = b/inputs = l and etc.) to solve this.
Is there any other better solution?delightful-barista-90363
06/30/2022, 8:46 PMblue-beach-27940
07/01/2022, 2:28 AMblue-beach-27940
07/01/2022, 3:45 AMblue-beach-27940
07/01/2022, 3:45 AMgray-architect-29447
07/01/2022, 5:38 AMlate-bear-87552
07/01/2022, 7:54 AM'typing-extensions-4.2.0 typing-inspect-0.7.1 tzdata-2022.1 tzlocal-4.2 urllib3-1.26.9 websocket-client-1.3.3 wrapt-1.14.1\n'
'[2022-07-01 07:34:29,599] INFO {datahub.cli.ingest_cli:99} - DataHub CLI version: 0.8.38.4\n'
'[2022-07-01 07:34:30,173] INFO {datahub.cli.ingest_cli:115} - Starting metadata ingestion\n'
'[2022-07-01 07:44:41,592] INFO {datahub.ingestion.source.usage.bigquery_usage:975} - Starting log load from GCP Logging\n'
'/usr/local/bin/run_ingest.sh: line 26: 20382 Killed ( python3 -m datahub ingest -c "$4/$1.yml" )\n',
"2022-07-01 07:49:20.890195 [exec_id=cceb1680-1556-4712-bc57-0b1d631836a3] INFO: Failed to execute 'datahub ingest'",
'2022-07-01 07:49:20.890775 [exec_id=cceb1680-1556-4712-bc57-0b1d631836a3] INFO: Caught exception EXECUTING '
'task_id=cceb1680-1556-4712-bc57-0b1d631836a3, name=RUN_INGEST, stacktrace=Traceback (most recent call last):\n'
' File "/usr/local/lib/python3.9/site-packages/acryl/executor/execution/default_executor.py", line 121, in execute_task\n'
' self.event_loop.run_until_complete(task_future)\n'
' File "/usr/local/lib/python3.9/site-packages/nest_asyncio.py", line 89, in run_until_complete\n'
' return f.result()\n'
' File "/usr/local/lib/python3.9/asyncio/futures.py", line 201, in result\n'
' raise self._exception\n'
' File "/usr/local/lib/python3.9/asyncio/tasks.py", line 256, in __step\n'
' result = coro.send(None)\n'
' File "/usr/local/lib/python3.9/site-packages/acryl/executor/execution/sub_process_ingestion_task.py", line 115, in execute\n'
' raise TaskError("Failed to execute \'datahub ingest\'")\n'
"acryl.executor.execution.task.TaskError: Failed to execute 'datahub ingest'\n"]}
astonishing-dusk-99990
07/01/2022, 7:56 AMdatahub docker quickstart --quickstart-compose-file=docker-compose.quickstart.yml
but when I did the ingestion on postgres always N/A and after I checked on the container, it looks like on the container datahub-actions when I checked got an error. Anyone know?hallowed-dog-79615
07/01/2022, 8:41 AMproud-baker-56489
07/01/2022, 11:56 AMproud-baker-56489
07/01/2022, 11:58 AMproud-baker-56489
07/04/2022, 3:48 AM