Túlio Lima
08/25/2024, 4:21 AMKRISHIV GUBBA
08/25/2024, 7:27 AMRabea Yousef
08/25/2024, 9:36 AMHassan Razzaq
08/25/2024, 2:27 PMHassan Razzaq
08/25/2024, 2:28 PMHassan Razzaq
08/25/2024, 2:29 PMHassan Razzaq
08/25/2024, 2:54 PM{"status":422,"type":"<https://reference.airbyte.com/reference/errors#unprocessable-entity>","title":"unprocessable-entity","detail":"The body of the request was not understood","documentationUrl":null,"data":{"message":"json schema validation failed when comparing the data to the json schema. \nErrors: $: required property 'api_key' not found, $: required property 'url' not found "}}
This is the url I am using url = "http:localhost:8000/api/public/v1/sources"
Túlio Lima
08/25/2024, 3:51 PMTúlio Lima
08/25/2024, 5:34 PMCharles Bockelmann
08/25/2024, 5:38 PMurllib3.exceptions.MaxRetryError: HTTPConnectionPool(host='localhost', port=8000): Max retries exceeded with url: /api/public/v1/connections/sync (Caused by NewConnectionError('<urllib3.connection.HTTPConnection object at 0xffff8581b770>: Failed to establish a new connection: [Errno 111] Connection refused'))
After accessing the console inside the Airflow container and doing a curl on ports 80
and 8000
to localhost
, host.docker.internal
and the Airbyte container IP address, I always receive the same response:
<html>
<head><title>404 Not Found</title></head>
<body>
<center><h1>404 Not Found</h1></center>
<hr><center>nginx</center>
mike Trienis
08/25/2024, 5:55 PMQuang Nguyen
08/25/2024, 6:30 PMERROR sync-operations-4 i.a.i.b.d.t.TyperDeduperUtil(executeTypeAndDedupe):223 Encountered Exception on unsafe SQL for stream raw_postgres shipments with suffix , attempting with error handling com.google.cloud.bigquery.BigQueryException: Query error: Invalid datetime string "+20212-01-10T07:07:00.000000" at [2:1]
at com.google.cloud.bigquery.Job.reload(Job.java:424) ~[google-cloud-bigquery-2.37.0.jar:2.37.0]
at io.airbyte.integrations.destination.bigquery.typing_deduping.BigQueryDestinationHandler.execute(BigQueryDestinationHandler.kt:146) ~[io.airbyte.airbyte-integrations.connectors-destination-bigquery.jar:?]
at io.airbyte.integrations.base.destination.typing_deduping.TyperDeduperUtil.executeTypeAndDedupe(TyperDeduperUtil.kt:219) ~[airbyte-cdk-typing-deduping-0.41.4.jar:?]
at io.airbyte.integrations.destination.bigquery.operation.BigQueryStorageOperation.typeAndDedupe(BigQueryStorageOperation.kt:158) ~[io.airbyte.airbyte-integrations.connectors-destination-bigquery.jar:?]
at io.airbyte.integrations.base.destination.operation.AbstractStreamOperation.finalizeTable(AbstractStreamOperation.kt:315) ~[airbyte-cdk-typing-deduping-0.41.4.jar:?]
at io.airbyte.integrations.base.destination.operation.DefaultSyncOperation.finalizeStreams$lambda$9$lambda$8(DefaultSyncOperation.kt:138) ~[airbyte-cdk-typing-deduping-0.41.4.jar:?]
at java.base/java.util.concurrent.CompletableFuture$AsyncSupply.run(CompletableFuture.java:1768) ~[?:?]
at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1144) ~[?:?]
at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:642) ~[?:?]
at java.base/java.lang.Thread.run(Thread.java:1583) [?:?]
i'm using CDC wal mode. can someone help me this errorQuang Nguyen
08/25/2024, 6:42 PMJhonatas Kleinkauff
08/25/2024, 8:39 PMTúlio Lima
08/26/2024, 12:04 AMHoang Trung Hieu
08/26/2024, 12:30 AMSlackbot
08/26/2024, 4:36 AMDR
08/26/2024, 5:07 AM2024-08-26 04:36:41 INFO i.a.c.i.LineGobbler(voidCall):166 - ----- START DISCOVER -----
2024-08-26 04:36:41 INFO i.a.c.i.LineGobbler(voidCall):166 -
2024-08-26 04:36:58 INFO i.a.c.ConnectorWatcher(run):87 - Connector exited, processing output
2024-08-26 04:36:58 INFO i.a.c.ConnectorWatcher(run):90 - Output file jobOutput.json found
2024-08-26 04:36:58 INFO i.a.c.ConnectorWatcher(run):96 - Connector exited with 1
2024-08-26 04:36:58 INFO i.a.w.i.VersionedAirbyteStreamFactory(create):189 - Reading messages from protocol version 0.2.0
2024-08-26 04:36:58 WARN i.a.m.l.MetricClientFactory(getMetricClient):43 - MetricClient has not been initialized. Must call MetricClientFactory.CreateMetricClient before using MetricClient. Using a dummy client for now. Ignore this if Airbyte is configured to not publish any metrics.
2024-08-26 04:36:58 WARN i.a.w.i.VersionedAirbyteStreamFactory(internalLog):305 - Refusing to infer schema for 4975 files; using 10 files.
2024-08-26 04:36:58 WARN i.a.w.i.VersionedAirbyteStreamFactory(internalLog):305 - Refusing to infer schema for 3949 files; using 10 files.
2024-08-26 04:36:58 ERROR i.a.w.i.VersionedAirbyteStreamFactory(internalLog):304 - An error occurred inferring the schema.
Traceback (most recent call last):
File "/usr/local/lib/python3.10/site-packages/airbyte_cdk/sources/file_based/stream/default_file_based_stream.py", line 289, in _infer_file_schema
return await self.get_parser().infer_schema(self.config, file, self.stream_reader, self.logger)
File "/usr/local/lib/python3.10/site-packages/airbyte_cdk/sources/file_based/file_types/csv_parser.py", line 168, in infer_schema
for row in data_generator:
File "/usr/local/lib/python3.10/site-packages/airbyte_cdk/sources/file_based/file_types/csv_parser.py", line 55, in read_data
headers = self._get_headers(fp, config_format, dialect_name)
File "/usr/local/lib/python3.10/site-packages/airbyte_cdk/sources/file_based/file_types/csv_parser.py", line 110, in _get_headers
fp.seek(0)
io.UnsupportedOperation: underlying stream is not seekable
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/usr/local/lib/python3.10/site-packages/airbyte_cdk/sources/file_based/stream/default_file_based_stream.py", line 281, in _infer_schema
base_schema = merge_schemas(base_schema, task.result())
File "/usr/local/lib/python3.10/site-packages/airbyte_cdk/sources/file_based/stream/default_file_based_stream.py", line 291, in _infer_file_schema
raise SchemaInferenceError(
airbyte_cdk.sources.file_based.exceptions.SchemaInferenceError: Error inferring schema from files. Are the files valid? Contact Support if you need assistance.
file=<https://storage.googleapis.com/gp-install-stats/installs_com.test_202408_overview.csv>? format=filetype='csv' delimiter=',' quote_char='"' escape_char=None encoding='UTF16' double_quote=True null_values=set() strings_can_be_null=True skip_rows_before_header=0 skip_rows_after_header=0 header_definition=CsvHeaderFromCsv(header_definition_type='From CSV') true_values={'t', '1', 'on', 'yes', 'y', 'true'} false_values={'no', '0', 'f', 'false', 'n', 'off'} inference_type=<inferencetype.none:> ignore_errors_on_fields_mismatch=False stream=install_report
Traceback (most recent call last):
File "/usr/local/lib/python3.10/site-packages/airbyte_cdk/sources/file_based/stream/default_file_based_stream.py", line 289, in _infer_file_schema
return await self.get_parser().infer_schema(self.config, file, self.stream_reader, self.logger)
Jayant Kumar
08/26/2024, 6:42 AMJulie Choong
08/26/2024, 7:06 AMSlackbot
08/26/2024, 7:14 AMShubham
08/26/2024, 7:23 AMincremental sync
in the second case ?
Even if I go for a full table append
mode, How do I provide a changing end_date (I can't use current_Date because the source doesn't allow end_date-start_date
to be greater than 7Tom Montgomery
08/26/2024, 7:50 AMtags
field returns an object. Within this object there is a data
key whose value is an array of tags objects. This array is limited to 10 objects. Should the contact have more tags, there is another key, has_more
, which would be set to true
in the case of the contact having more than 10 tags. Finally, there is a URL provided within the tags
field that is used to get more resources for the contact (i.e., more tags). Would it be possible to automate the fetching of this additional data within the airbyte sync? At the moment we are only receiving 10 tags and this is obscuring our viewThomas
08/26/2024, 9:00 AMabctl local install --migrate
which seemingly worked but when trying to open the web interface I get a 404 error from nginxL Theisen
08/26/2024, 9:15 AMIshan Anilbhai Koradiya
08/26/2024, 9:45 AMSyed Hamza Raza Kazmi
08/26/2024, 9:53 AMQuang Nguyen
08/26/2024, 9:55 AMStack Trace: com.google.cloud.bigquery.BigQueryException: Query error: Invalid NUMERIC value: 17976931348623157000000000000000... at [2:1]
at com.google.cloud.bigquery.Job.reload(Job.java:424)
at io.airbyte.integrations.destination.bigquery.typing_deduping.BigQueryDestinationHandler.execute(BigQueryDestinationHandler.kt:146)
at io.airbyte.integrations.base.destination.typing_deduping.TyperDeduperUtil.executeTypeAndDedupe(TyperDeduperUtil.kt:219)
at io.airbyte.integrations.destination.bigquery.operation.BigQueryStorageOperation.typeAndDedupe(BigQueryStorageOperation.kt:158)
at io.airbyte.integrations.base.destination.operation.AbstractStreamOperation.finalizeTable(AbstractStreamOperation.kt:315)
at io.airbyte.integrations.base.destination.operation.DefaultSyncOperation.finalizeStreams$lambda$9$lambda$8(DefaultSyncOperation.kt:138)
at java.base/java.util.concurrent.CompletableFuture$AsyncSupply.run(CompletableFuture.java:1768)
at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1144)
at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:642)
at java.base/java.lang.Thread.run(Thread.java:1583)
Aditya Gupta
08/26/2024, 10:34 AMSaurabh Agrawal
08/26/2024, 11:21 AMSYNC_JOB_MAX_ATTEMPTS
in the Airbyte Cloud accont..