Robert Put
10/12/2022, 10:29 PMwp
10/12/2022, 10:37 PMBigQueryError{reason=invalidQuery, location=query, message=Field ad_group_criterion_final_urls has incompatible types. Configured schema: string. Avro file: record}
Is this a bug on the source connector?Mohit Reddy
10/13/2022, 3:48 AM2022-10-13 03:37:17 INFO i.a.w.p.KubePodProcess(<init>):572 - Pod IP: 10.105.236.201
2022-10-13 03:37:17 INFO i.a.w.p.KubePodProcess(<init>):579 - Using null stdin output stream...
2022-10-13 03:37:17 ERROR i.a.w.g.DefaultCheckConnectionWorker(run):98 - Unexpected error while checking connection:
java.lang.NullPointerException: null
at java.io.Reader.<init>(Reader.java:168) ~[?:?]
at java.io.InputStreamReader.<init>(InputStreamReader.java:112) ~[?:?]
at io.airbyte.commons.io.IOs.newBufferedReader(IOs.java:120) ~[io.airbyte-airbyte-commons-0.40.14.jar:?]
at io.airbyte.commons.io.LineGobbler.<init>(LineGobbler.java:99) ~[io.airbyte-airbyte-commons-0.40.14.jar:?]
at io.airbyte.commons.io.LineGobbler.gobble(LineGobbler.java:67) ~[io.airbyte-airbyte-commons-0.40.14.jar:?]
at io.airbyte.commons.io.LineGobbler.gobble(LineGobbler.java:28) ~[io.airbyte-airbyte-commons-0.40.14.jar:?]
at io.airbyte.workers.general.DefaultCheckConnectionWorker.run(DefaultCheckConnectionWorker.java:65) ~[io.airbyte-airbyte-workers-0.40.14.jar:?]
at io.airbyte.workers.general.DefaultCheckConnectionWorker.run(DefaultCheckConnectionWorker.java:36) ~[io.airbyte-airbyte-workers-0.40.14.jar:?]
at io.airbyte.workers.temporal.TemporalAttemptExecution.lambda$getWorkerThread$2(TemporalAttemptExecution.java:161) ~[io.airbyte-airbyte-workers-0.40.14.jar
To get around this, we disable injecting the sidecar by adding an annotation to the jobs - https://docs.airbyte.com/operator-guides/configuring-airbyte/#jobs-specific (specifically CHECK_JOB_KUBE_ANNOTATIONS). We recently upgraded airbyte from 0.39.1 to 0.40.14 and this has started to fail again i.e. the annotation is not being applied. Any help here?Shashank Tiwari
10/13/2022, 10:22 AM0/2 nodes are available: 1 Too many pods, 1 node(s) didn't find available persistent volumes to bind.
This is the warning coming in the pod for airbyte-db when deployed on EKS. Can anyone help me in thisSefath Chowdhury
10/13/2022, 12:42 PMsonti srihari
10/13/2022, 12:48 PMNir Chamo
10/13/2022, 1:32 PMStuart Horgan
10/13/2022, 3:46 PMdef read_records(self, *args, **kwargs) -> Iterable[Mapping[str, Any]]:
for record in super().read_records(*args, **kwargs):
if self._cursor_value:
latest_record_date = record[self.cursor_field]
self._cursor_value = max(self._cursor_value, latest_record_date)
yield record
and this is how we get the new state value for future runs to use. But the first time you run it, the self._cursor_value is set to None, so we never enter the if statement and update to the latest value. So what is supposed to happen here? How do we get the correct state returned at the end of the first run instead of the start date being returned unchanged?Alberto Colon
10/13/2022, 2:31 PMrce-postgres-check-9688c843-03b2-40af-9b99-98b9d45e18f9-0-ixmtc 0/4 Init:Error 0 6m58s 10.42.7.8 b020114-47b9054 <none> <none>
The logs of this pod:
Timeout while attempting to copy to init container, exiting with code 1...
and in the airbyte UI we get this error (attachment):
Do I need to setup something else in the env file for kubernetes? I really don't have any clue about this bug/error.
Thanks in advance...Jhon Edison Bambague Calderon
10/13/2022, 3:33 PMJordan Young
10/13/2022, 4:39 PMsar
10/13/2022, 5:58 PMdestination-aws-datalake
connector. We have a table that’s more than 15+ million rows and whenever we try to run a sync using that connector, the docker container that gets spun out to run the sync ends up chewing all the host memory and eventually crashing the EC2 instance. Tried setting some global docker limits (as i can’t do it on the container level since it gets spun up when the sync job starts) to no success as docker stats
kept showing the available memory on the host. Tried resizing the instance and even with 64GB of RAM on the host, still ran into the same issue.
https://github.com/airbytehq/airbyte/tree/master/airbyte-integrations/connectors/destination-aws-datalake
Anyone run into something similar using that connector and how did you end up resolving it?
Here’s a bit more info on the issue itself - https://github.com/mitodl/ol-data-platform/issues/371Dusty Shapiro
10/13/2022, 6:34 PMpostgresql:
enabled: false
externalDatabase:
host: ${db_host}
user: ${db_user}
password: ${db_password}
database: ${db_name}
port: ${db_port}
Pulkit Srivastava
10/13/2022, 7:15 PMsync
endpoint is taking more time than when we were using 0.33.5-alpha can you please let me know if this is expected behaviour or not previously it was taking less than 1 sec in new version its taking 6 to 7 seconds. If any more info is needed I am happy to provide. ThanksDavid hatch
10/13/2022, 8:04 PMis of type timestamp without time zone but expression is of type text
. In my schema i have configured this column as follows:
"block_day": {
"type": "string",
"format": "date-time",
"airbyte_type": "timestamp_without_timezone"
},
Any ideas on how to resolve this would be appreciated. I’ve tried a couple variations of the configuration and have searched around on the airbyte github issues but haven’t found a solution.Matt Webster
10/13/2022, 8:07 PM13 of 33 ERROR creating incremental model loading.log_scd................................................... [ERROR in 16.07s]
Database Error in model log_scd (models/generated/airbyte_incremental/scd/loading/log_scd.sql)
Invalid input
DETAIL:
-----------------------------------------------
error: Invalid input
code: 8001
context: CONCAT() result too long for type varchar(65535)
query: 417552
location: string_ops.cpp:108
process: query0_113_417552 [pid=690]
-----------------------------------------------
compiled SQL at ../build/run/airbyte_utils/models/generated/airbyte_incremental/scd/loading/log_scd.sql
Any ideas would be greatly appreciated!
Here is the root issue in GitHub: https://github.com/airbytehq/airbyte/issues/14441Eduardo Aviles
10/13/2022, 6:52 PM2022-10-12 09:22:30 [44msource[0m > Encountered an exception while reading stream items
Traceback (most recent call last):
File "/usr/local/lib/python3.9/site-packages/requests/models.py", line 971, in json
return complexjson.loads(self.text, **kwargs)
File "/usr/local/lib/python3.9/json/__init__.py", line 346, in loads
return _default_decoder.decode(s)
File "/usr/local/lib/python3.9/json/decoder.py", line 337, in decode
obj, end = self.raw_decode(s, idx=_w(s, 0).end())
File "/usr/local/lib/python3.9/json/decoder.py", line 355, in raw_decode
raise JSONDecodeError("Expecting value", s, err.value) from None
json.decoder.JSONDecodeError: Expecting value: line 1 column 1 (char 0)
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/usr/local/lib/python3.9/site-packages/airbyte_cdk/sources/abstract_source.py", line 115, in read
yield from self._read_stream(
File "/usr/local/lib/python3.9/site-packages/airbyte_cdk/sources/abstract_source.py", line 165, in _read_stream
for record in record_iterator:
File "/usr/local/lib/python3.9/site-packages/airbyte_cdk/sources/abstract_source.py", line 254, in _read_full_refresh
for record in records:
File "/usr/local/lib/python3.9/site-packages/airbyte_cdk/sources/streams/http/http.py", line 417, in read_records
response = self._send_request(request, request_kwargs)
File "/usr/local/lib/python3.9/site-packages/airbyte_cdk/sources/streams/http/http.py", line 336, in _send_request
return backoff_handler(user_backoff_handler)(request, request_kwargs)
File "/usr/local/lib/python3.9/site-packages/backoff/_sync.py", line 105, in retry
ret = target(*args, **kwargs)
File "/usr/local/lib/python3.9/site-packages/backoff/_sync.py", line 105, in retry
ret = target(*args, **kwargs)
File "/usr/local/lib/python3.9/site-packages/airbyte_cdk/sources/streams/http/http.py", line 291, in _send
if self.should_retry(response):
File "/airbyte/integration_code/source_monday/source.py", line 55, in should_retry
is_complex_query = response.json().get("errors")
File "/usr/local/lib/python3.9/site-packages/requests/models.py", line 975, in json
raise RequestsJSONDecodeError(e.msg, e.doc, e.pos)
requests.exceptions.JSONDecodeError: Expecting value: line 1 column 1 (char 0)
2022-10-12 09:22:30 [44msource[0m > Finished syncing items
Simon Thelin
10/13/2022, 8:43 PMupdate
column on a pg
table, it seems to read through all data of the table, but only writes the delta
changes.
On some table syncs however, I can see it only reads very small chunks.
Is there a particular reason for this behaviour?Luca Moity
10/13/2022, 1:42 PMTammy Shipps
10/13/2022, 10:04 PMEdgar Valdez
10/14/2022, 2:41 AMpublic
folder? ie:
Target bucket: abc
and after running the job, tables are stored at: abc/public/table_1
I’d like: abc/table_1
Cheers!
PS: Target file is a parquet fileWayne
10/14/2022, 3:24 AMManuveeran Sandhu
10/14/2022, 6:12 AMBogdan
10/13/2022, 1:45 PMEmilja Dankevičiūtė
10/14/2022, 9:43 AMlaila ribke
10/14/2022, 11:52 AMTravis James
10/14/2022, 12:43 PMKevin Millan
10/10/2022, 9:31 PMIncremental | Deduped + history
table by migrating from redshift to postgres but, I encountered a couple of errors along the way. I'm using a datetime
column as the cursor field
:
1- I was first getting an error from airbyte saying that the table *_stg
needed a REPLICA IDENTITY
. I then added the field manually with ALTER TABLE analytics.revenue REPLICA IDENTITY FULL;
2- Now I'm getting the same error (log file attached) but with the *_scd
table. But the issue seems to be that it is having problems created the _scd
table in the first place.Robert Put
10/14/2022, 6:39 PMLucas Gonthier
10/14/2022, 6:50 PM