Thomas Pedot
01/23/2023, 1:07 PMMessage: Failed to inject value for parameter [localDockerMount] of method [checkDockerProcessFactory] of class: io.airbyte.workers.process.ProcessFactory
Message: Error resolving property value [${airbyte.local.docker-mount}]. Property doesn't exist
I feel that it is it related to my k3s setup ? I don't see where localDockerMount can be set ?
For the context : I am trying to create a custom connector with a gitlab private registry. I have setup my secret to connect to it. I don't know yet if it is good (and not sure it is related to this error either).Rishabh Jain
01/23/2023, 2:08 PMThomas Pedot
01/23/2023, 2:11 PMversion: "0.1.0"
definitions:
selector:
extractor:
field_pointer: []
requester:
url_base: "<https://axonaut.com/api/v2>"
http_method: "GET"
authenticator:
type: ApiKeyAuthenticator
header: "userApiKey"
api_token: "{{ config['userApiKey'] }}"
request_options_provider:
request_parameters:
date_after: "{{ config['date_after'] }}"
stream_slicer:
type: "DatetimeStreamSlicer"
start_datetime:
datetime: "{{ config['date_after'] }}"
datetime_format: "%d/%m/%Y"
end_datetime:
datetime: "{{ now_utc() }}"
datetime_format: "%Y-%m-%d %H:%M:%S.%f+00:00"
step: "P1D"
datetime_format: "%d/%m/%Y"
cursor_field: "{{ options['stream_cursor_field'] }}"
cursor_granularity: "P1D"
retriever:
record_selector:
$ref: "*ref(definitions.selector)"
paginator:
type: NoPagination
requester:
$ref: "*ref(definitions.requester)"
stream_slicer:
$ref: "*ref(definitions.stream_slicer)"
base_stream:
retriever:
$ref: "*ref(definitions.retriever)"
invoices_stream:
$ref: "*ref(definitions.base_stream)"
$options:
name: "invoices"
primary_key: "id"
path: "/invoices"
request: ""
stream_cursor_field: "date"
streams:
- "*ref(definitions.invoices_stream)"
check:
stream_names:
- "invoices"
spec:
documentation_url: <https://docs.airbyte.com/integrations/sources/axonaut>
connection_specification:
title: Axonaut Spec
type: object
required:
- userApiKey
additionalProperties: true
properties:
# 'TODO: This schema defines the configuration required for the source. This usually involves metadata such as database and/or authentication information.':
userApiKey:
type: string
description: API access key used to retrieve data from the Exchange Rates API.
airbyte_secret: true
date_after:
type: string
description: Start getting data from that date.
pattern: ^(0[1-9]|[12][0-9]|3[01])[- /.](0[1-9]|1[012])[- /.](19|20)\d\d$
examples:
- "%d/%m/%Y"
Roy Ben Dov
01/23/2023, 3:12 PMThe connection tests failed.
Could not connect with provided configuration. Error: Expected exactly one replication slot but found 0. Please read the docs and add a replication slot to your database.
I followed through the documentation and a replication slot exists in the dbAlexandre Keo
01/23/2023, 3:48 PMOmer Kolodny
01/23/2023, 4:18 PM2023-01-23 12:39:23 INFO i.a.w.g.DefaultReplicationWorker(getReplicationOutput):482 - failures: [ {
"failureOrigin" : "destination",
"internalMessage" : "io.airbyte.workers.general.DefaultReplicationWorker$DestinationException: Destination process message delivery failed",
"externalMessage" : "Something went wrong within the destination connector",
Any idea what is this error?Peter Kong
01/23/2023, 4:56 PMnormalized tabular data
is toggled; tables are fully synced.
2. I edit the Replication
settings and remove a stream from my connection
3. A few minutes later, observe that Raw data (JSON)
is toggled in the Transformation tab.
This is concerning because I don't know the state of my data (whether it's normalized or not), or why the transformation setting has reverted to Raw data (JSON)
. Thank you for your help.Hassan Shahid
01/23/2023, 5:28 PMTRUNCATECOLUMNS
in the list of redshift COPY options). I’ve got the custom image pushed up to our private registry and added as a connector to our airbyte instance. When I use it, it fails with this error:
Caused by: java.lang.IllegalStateException: Requested normalization for <http://xxx.dkr.ecr.xxx.amazonaws.com/xxx/destination-redshift:0.1.0|xxx.dkr.ecr.xxx.amazonaws.com/xxx/destination-redshift:0.1.0>, but it is not included in the normalization mappings.
at io.airbyte.workers.normalization.NormalizationRunnerFactory.getNormalizationInfoForConnector(NormalizationRunnerFactory.java:57) ~[io.airbyte-airbyte-workers-0.39.28-alpha.jar:?]
at io.airbyte.workers.normalization.NormalizationRunnerFactory.create(NormalizationRunnerFactory.java:43) ~[io.airbyte-airbyte-workers-0.39.28-alpha.jar:?]
at io.airbyte.workers.temporal.sync.NormalizationActivityImpl.lambda$getLegacyWorkerFactory$5(NormalizationActivityImpl.java:115) ~[io.airbyte-airbyte-workers-0.39.28-alpha.jar:?]
at io.airbyte.workers.temporal.TemporalAttemptExecution.get(TemporalAttemptExecution.java:118) ~[io.airbyte-airbyte-workers-0.39.28-alpha.jar:?]
at io.airbyte.workers.temporal.sync.NormalizationActivityImpl.lambda$normalize$3(NormalizationActivityImpl.java:103) ~[io.airbyte-airbyte-workers-0.39.28-alpha.jar:?]
at io.airbyte.workers.temporal.TemporalUtils.withBackgroundHeartbeat(TemporalUtils.java:284) ~[io.airbyte-airbyte-workers-0.39.28-alpha.jar:?]
I’ve looked through the forum, stack overflow, google. and searched through this slack, but can’t find any mention of this failure. Can anyone point me in the right direction?Massy Bourennani
01/23/2023, 6:32 PMairbyte_emitted_at
column
2. new rows according to a cursor_column
most of the time it’s an updated_at
(a column that airbyte would typically use to sync streams incrementally)
wanted to know how everyone was tackling this. Example in thread 🧵Andres Gutierrez
01/23/2023, 7:56 PMNO
but I wanted to ask if is possible in some way to do casting of Sheet's columns with this connector?
https://docs.airbyte.com/integrations/sources/google-sheets/#data-type-mappingWalker Philips
01/23/2023, 8:05 PMChen Lin
01/23/2023, 9:35 PMdisplay_keyword_performance_report
stream data, I don't see any obvious errors from the log but the file that's written in s3 bucket is empty, I don't have any issue with other streams like ad_group_ad_report or campaigns, attached is the log file, can you guys suggest where I should look into? Thanks!Adrian Bakula
01/23/2023, 10:22 PMAmit Khanna
01/23/2023, 10:47 PMJason Maddern
01/23/2023, 11:16 PM_airbyte_ab_id
, but I cannot see how to join on that key.
Can anyone advise how to join normalised JSON post airbyte run? I must be missing something obvious
I'm using airbyte latest, connected to snowflakeAbhinav Kapur
01/23/2023, 11:25 PMNathan Chan
01/24/2023, 1:11 AMJoey Taleño
01/24/2023, 4:17 AMShreshth Arora
01/24/2023, 6:11 AMJOB_MAIN_CONTAINER_MEMORY_REQUEST=12g
JOB_MAIN_CONTAINER_MEMORY_LIMIT= 12g
JOB_MAIN_CONTAINER_CPU_REQUEST= 0.75
JOB_MAIN_CONTAINER_CPU_LIMIT= 0.80
MAX_SYNC_WORKERS=20
MAX_SPEC_WORKERS=20
MAX_CHECK_WORKERS=20
MAX_DISCOVER_WORKERS=20
MAX_NOTIFY_WORKERS=20
Similar Configs for Normalisation Container Variables also. We have run multiple syncs of multiple sizes but haven’t been able to increase the ETA of sync and is constant. Usage of the machine from airbyte to isn’t more than 40% of CPU and 4GBs of RAM (out of 32GB).
With current test sync speeds Airbyte would take days to sync data for our needs (BigQuery to Clickhouse)
1) Anyway to configure the system in the best possible way for full utilization of the machine ?
2) What are the best practices for scaling Airbyte docker deployment to its full potential?Suprakash Nandy
01/24/2023, 6:35 AMRoy Ben Dov
01/24/2023, 8:54 AMCould not connect with provided SSH configuration. Error: org.apache.sshd.common.SshException: No more authentication methods available
can someone elaborate ?Grember Yohan
01/24/2023, 8:57 AMreact-dom.production.min.js:216 Error: Internal Server Error: Cannot invoke "io.airbyte.config.storage.CloudStorageConfigs.getType()" because the return value of "io.airbyte.config.helpers.LogConfigs.getStorageConfigs()" is null
at apiOverride.ts:107:9
at f (regeneratorRuntime.js:86:17)
at Generator._invoke (regeneratorRuntime.js:66:24)
at Generator.next (regeneratorRuntime.js:117:21)
at r (asyncToGenerator.js:3:20)
at u (asyncToGenerator.js:25:9)
Downgrading from 0.40.28 to 0.40.27 fixes this issue.
Should I share this somewhere specific to document this regression and improve its chances to be fixed?Keurcien Luu
01/24/2023, 10:23 AMdef get_items(self, start_date: datetime):
return self._gateway.transaction.search(braintree.TransactionSearch.created_at >= start_date)
Maybe we can iterate over the dates since start_date to get more records (in our case 50k per day would be largely enough), but we could run into some rate limiting issues I guess. Any advice on that?Danilo Drobac
01/24/2023, 10:49 AMAndrzej Lewandowski
01/24/2023, 10:49 AMShraddha Borkar
01/24/2023, 10:50 AMCould not connect to the Kafka brokers with provided configuration. Failed to construct kafka producer
Could someone please help here?Sharath Chandra
01/24/2023, 11:03 AM(models/generated/airbyte_tables/airbyte_test/<table_name>.sql)
Invalid input
DETAIL:
-----------------------------------------------
error: Invalid input
code: 8001
context: CONCAT() result too long for type varchar(65535)
query: 19684211
location: string_ops.cpp:108
process: query0_121_19684211 [pid=17065]
Omer Kolodny
01/24/2023, 1:32 PMAmit Khanna
01/24/2023, 1:36 PMIgor Safonov
01/24/2023, 1:44 PM