billowy-rocket-47022
03/10/2022, 5:37 PMlemon-terabyte-66903
03/10/2022, 5:55 PMmysterious-australia-30101
03/11/2022, 8:25 AMprehistoric-optician-40107
03/16/2022, 1:34 PMdatahub-gms:8080
and got error again.
OperationalError: (psycopg2.OperationalError) connection to server at "localhost" (127.0.0.1), port 5432 failed: Connection refused\n'
'\tIs the server running on that host and accepting TCP/IP connections?\n'
'connection to server at "localhost" (::1), port 5432 failed: Cannot assign requested address\n'
'\tIs the server running on that host and accepting TCP/IP connections?\n'
'\n'
'(Background on this error at: <http://sqlalche.me/e/13/e3q8>)\n',
"2022-03-16 13:05:46.799054 [exec_id=980fe2a1-816d-4970-95ec-09727322c3ea] INFO: Failed to execute 'datahub ingest'",
'2022-03-16 13:05:46.799641 [exec_id=980fe2a1-816d-4970-95ec-09727322c3ea] INFO: Caught exception EXECUTING '
'task_id=980fe2a1-816d-4970-95ec-09727322c3ea, name=RUN_INGEST, stacktrace=Traceback (most recent call last):\n'
' File "/usr/local/lib/python3.9/site-packages/acryl/executor/execution/default_executor.py", line 119, in execute_task\n'
' self.event_loop.run_until_complete(task_future)\n'
' File "/usr/local/lib/python3.9/site-packages/nest_asyncio.py", line 81, in run_until_complete\n'
' return f.result()\n'
' File "/usr/local/lib/python3.9/asyncio/futures.py", line 201, in result\n'
' raise self._exception\n'
' File "/usr/local/lib/python3.9/asyncio/tasks.py", line 256, in __step\n'
' result = coro.send(None)\n'
' File "/usr/local/lib/python3.9/site-packages/acryl/executor/execution/sub_process_ingestion_task.py", line 115, in execute\n'
' raise TaskError("Failed to execute \'datahub ingest\'")\n'
"acryl.executor.execution.task.TaskError: Failed to execute 'datahub ingest'\n"]}
Execution finished with errors.
I can connect normally to PostgreSQL, I couldn't figure out why it gave such an error.billowy-book-26360
03/17/2022, 1:23 AMred-napkin-59945
03/24/2022, 11:59 PMnumerous-cricket-19689
03/29/2022, 12:02 AMbitter-toddler-42943
03/29/2022, 2:15 AMbitter-toddler-42943
03/29/2022, 2:17 AMpip install 'acryl-datahub[datahub-rest]'
already , is there anything else that I have to check more?bitter-toddler-42943
03/29/2022, 2:19 AMfresh-memory-10355
03/29/2022, 4:15 PMaverage-france-59117
03/30/2022, 8:37 AMParserError
while ingesting BigQuery profiling. This error occurs only when profiling monthly partitioned table. The error message is following:
...
File "/Users/seb.kim/github/datahub/dh/lib/python3.8/site-packages/datahub/ingestion/run/pipeline.py", line 181, in run
for wu in itertools.islice(
File "/Users/seb.kim/github/datahub/dh/lib/python3.8/site-packages/datahub/ingestion/source/sql/bigquery.py", line 655, in get_workunits
for wu in super().get_workunits():
File "/Users/seb.kim/github/datahub/dh/lib/python3.8/site-packages/datahub/ingestion/source/sql/sql_common.py", line 656, in get_workunits
profile_requests += list(
File "/Users/seb.kim/github/datahub/dh/lib/python3.8/site-packages/datahub/ingestion/source/sql/sql_common.py", line 1157, in loop_profiler_requests
(partition, custom_sql) = self.generate_partition_profiler_query(
File "/Users/seb.kim/github/datahub/dh/lib/python3.8/site-packages/datahub/ingestion/source/sql/bigquery.py", line 585, in generate_partition_profiler_query
partition_datetime = parser.parse(partition.partition_id)
File "/Users/seb.kim/github/datahub/dh/lib/python3.8/site-packages/dateutil/parser/_parser.py", line 1368, in parse
return DEFAULTPARSER.parse(timestr, **kwargs)
File "/Users/seb.kim/github/datahub/dh/lib/python3.8/site-packages/dateutil/parser/_parser.py", line 651, in parse
six.raise_from(ParserError(str(e) + ": %s", timestr), e)
File "<string>", line 3, in raise_from
ParserError: month must be in 1..12: 202203
When I ingest daily
partitioned table like the first, there’s no issue at all. However, for the monthly
partitioned table like the second, it emits ParserError
. FYI, for the first figure, timestamp TYPE is TIMESTAMP
and for the second figure record_date TYPE is DATE
in BigQuery. Any advice??quick-engine-33953
03/31/2022, 1:16 PMsource:
type: bigquery
config:
project_id: awesome-project
env: prod
credential:
project_id: awesome-project
private_key_id: REDACTED
private_key: "REDACTED"
client_email: REDACTED
client_id: 'REDACTED'
include_tables: true
include_views: true
include_table_lineage: true
start_time: '2022-01-01T00:00:00.000Z'
end_time: '2022-12-31T00:00:00.000Z'
use_exported_bigquery_audit_metadata: true
profile_pattern:
allow:
- schema.table.column
deny:
- '*.*.*'
sink:
type: datahub-rest
config:
server: '<http://REDACTED:8080>'
All the permissions in Google Cloud's roles are configured as well.red-smartphone-15526
03/31/2022, 1:20 PMRetryError: HTTPSConnectionPool(host='<http://redashserver.org|redashserver.org>', port=443): Max retries exceeded with url: /api/dashboards/overview (Caused by ResponseError('too many 500 error responses'))
Has anyone faced similar issues?nutritious-bird-77396
03/31/2022, 9:08 PMDatahubEmitterOperator.
It doesn't work with with bashOperator
As a result only the dataset lineage is sent and not the Pipeline/Tasks info...
Do you have any insights on this?fresh-memory-10355
04/01/2022, 6:47 AMincalculable-apartment-22203
04/01/2022, 12:49 PMincalculable-apartment-22203
04/01/2022, 1:11 PMbetter-orange-49102
04/05/2022, 9:42 AMstocky-midnight-78204
04/05/2022, 11:23 AMbrave-forest-5974
04/05/2022, 1:26 PMred-napkin-59945
04/07/2022, 4:48 PMbitter-toddler-42943
04/12/2022, 1:37 AMbitter-toddler-42943
04/12/2022, 6:23 AMcold-hydrogen-10513
04/12/2022, 9:15 AMswift-breakfast-25077
04/12/2022, 6:33 PMcurved-crayon-1929
04/12/2022, 7:54 PMcreamy-van-28626
04/13/2022, 6:00 AMcreamy-van-28626
04/14/2022, 7:57 AMfamous-match-44342
04/14/2022, 1:32 PM