Krishna Channa
08/28/2024, 4:05 PMAditya Gupta
08/28/2024, 4:14 PMGordon MacMaster
08/28/2024, 4:30 PMInternal message: com.mongodb.MongoCommandException: Command failed with error 292 (QueryExceededMemoryLimitNoDiskUseAllowed): 'PlanExecutor error during aggregation :: caused by :: Sort exceeded memory limit of 33554432 bytes, but did not opt in to external sorting. Aborting operation. Pass allowDiskUse:true to opt in.' on server {REMOVED}. The full response is {"ok": 0.0, "errmsg": "PlanExecutor error during aggregation :: caused by :: Sort exceeded memory limit of 33554432 bytes, but did not opt in to external sorting. Aborting operation. Pass allowDiskUse:true to opt in.", "code": 292, "codeName": "QueryExceededMemoryLimitNoDiskUseAllowed", "$clusterTime": {"clusterTime": {"$timestamp": {"t": 1724862367, "i": 4}}, "signature": {"hash": {"$binary": {"base64": "{REMOVED}", "subType": "00"}}, "keyId": {REMOVED}}}, "operationTime": {"$timestamp": {"t": 1724862367, "i": 4}}}
Narayan Zeermire
08/28/2024, 6:35 PMBrian Kasen
08/28/2024, 7:10 PM2024-08-28 12:35:49 [43mdestination[0m > INFO main i.a.c.d.j.JdbcDatabase(executeWithinTransaction$lambda$1):46 executing query within transaction: insert into "airbyte_internal"."_airbyte_destination_state" ("name", "namespace", "destination_state", "updated_at") values ('airbyte_brand', 'TALENTREEF', '{"needsSoftReset":false,"airbyteMetaPresentInRaw":true}', '2024-08-28T12:35:48.549790446Z')
, but shortly thereafter I see :
2024-08-28 12:35:51 [43mdestination[0m > INFO sync-operations-3 i.a.i.b.d.t.TyperDeduperUtil(executeTypeAndDedupe):212 Attempting typing and deduping for TALENTREEF.airbyte_brand with suffix _ab_soft_reset
We have syncs that run every 8 hours and, what I suspect is a bug, is quickly becoming cost prohibitive since the long-term solution is not sizing up the Snowflake WH.
Kapa did not provide sufficient info on why SoftReset was triggering in this case
Has anyone encountered this before and can anyone from the Airbyte Team assist?
cc @Abhra Gupta / @Ritika NaiduSumit Kumar
08/28/2024, 7:23 PMBeatrice Nasike
08/28/2024, 7:52 PMEric Markewitz
08/28/2024, 8:08 PMuname -m
aarch64
curl -LsfS <https://get.airbyte.com> | bash -
Installing for Linux...
Downloading abctl from <https://github.com/airbytehq/abctl/releases/download/v0.13.1/abctl-v0.13.1-linux-amd64.tar.gz>
Installing 'release/abctl-v0.13.1-linux-amd64/abctl' to /usr/local/bin
bash: line 242: /usr/local/bin/abctl: cannot execute binary file
abctl install failed: bash: line 242: /usr/local/bin/abctl: cannot execute binary file
Herbert Sousa
08/28/2024, 9:42 PM{
"jobId": 17013431,
"status": "succeeded",
"jobType": "sync",
"startTime": "2024-08-28T18:01:51Z",
"connectionId": "223b217d-8398-4920-b979-1dc8d5a28ec5",
"lastUpdatedAt": "2024-08-28T18:10:53Z",
"duration": "PT9M2S",
"bytesSynced": 828539,
"rowsSynced": 4512
}
Sean Miltenberger
08/28/2024, 10:40 PMKrishna Channa
08/29/2024, 12:50 AMJohn Dorlus (Power Coder)
08/29/2024, 12:50 AMDean Lau
08/29/2024, 6:38 AMretain
OR remove
record?Lisardo Erman
08/29/2024, 7:09 AMYannis Thomopoulos
08/29/2024, 8:37 AMAditya Gupta
08/29/2024, 8:38 AMAlexandre Martins
08/29/2024, 9:36 AMsource_config.json
in the local filesystem, during the check process. However, as I run Airbyte OSS on Kubernetes (EKS), I'm getting a permission denied when the job pod tries to create this file:
PermissionError: [Errno 13] Permission denied: 'source_config.json'
with open(config_path, "w") as fh:
File "/usr/local/lib/python3.10/site-packages/airbyte_cdk/connector.py", line 60, in write_config
source.write_config(migrated_config, config_path)
File "/airbyte/integration_code/source_facebook_marketing/config_migrations.py", line 186, in _modify_and_save
cls._modify_and_save(config_path, source, config),
File "/airbyte/integration_code/source_facebook_marketing/config_migrations.py", line 160, in migrate
MigrateSecretsPathInConnector.migrate(sys.argv[1:], source)
File "/airbyte/integration_code/source_facebook_marketing/run.py", line 18, in run
run()
File "/airbyte/integration_code/main.py", line 9, in <module>
Traceback (most recent call last):
I suspect it has to do with the securityContext of the job pods, not allowing for write access to the internal filesystem. However, I don't see how we can update the security context of the job/check pods in the charts, only annotations, labels, etc. :https://github.com/airbytehq/airbyte-platform/blob/4aa1fd563b22802d268febfc5f61bbc928c40b33/charts/airbyte/values.yaml#L136-L170
Any ideas on how to solve this? Its blocking us from upgrading to the latest version 😕user
08/29/2024, 9:44 AMglobal.storage.gcs.credentialsJson
.
However it is not possible, as the Helm chart always requires it.
The remedy is to update the Helm chart to allow global.storage.gcs.credentialsJson
to be empty.
airbytehq/airbyteJulien Ruey
08/29/2024, 10:10 AMuser
08/29/2024, 10:13 AMLisardo Erman
08/29/2024, 10:19 AMuser
08/29/2024, 10:26 AMuser
08/29/2024, 11:07 AMplatform > Schema validation errors found for stream _GET_VENDOR_INVENTORY_REPORT. Error messages: [$.startDate: 2024-05-01 is an invalid date-time, $.endDate: 2024-05-31 is an invalid date-time]
Dates are set in the propper format. Please help.
airbytehq/airbyteSANGADO
08/29/2024, 11:36 AMdhanesh
08/29/2024, 12:05 PMWillian Yoshio Iwamoto
08/29/2024, 12:13 PMKornel All
08/29/2024, 12:31 PMStockton Fisher
08/29/2024, 12:54 PM> message='io.airbyte.workers.exception.WorkloadMonitorException: Airbyte could not track the sync progress. No heartbeat within the time limit indicates the process might have died.', type='java.lang.RuntimeException', nonRetryable=false
user
08/29/2024, 1:25 PMvalues.yaml
file as mentioned in the documentation
global:
storage:
type: "GCS"
storageSecretName: airbyte-config-secrets
bucket:
log: airbyte-bucket
state: airbyte-bucket
workloadOutput: airbyte-bucket
gcs:
projectId: <project-id>
Later, during the deployment, this secret is used to populate the gcs-log-creds-volume
volume for at least the airbyte-workload-api-server
and airbyte-server
charts
Issue
While this is working for airbyte-server
charts I've found an issue w/ airbyte-workload-api-server
.
The airbyte-workload-api-server
deployment fails because of this err :
At the end, the secret you've provided in your values.yaml is not used, the default one is kept instead Code InMountVolume.SetUp failed for volume "gcs-log-creds-volume" : secret "airbyte-gcs-log-creds" not found
airbyte-server
chart deployment file
volumes:
{{- if eq .Values.global.deploymentMode "oss" }}
{{- if eq (lower (default "" .Values.global.storage.type)) "gcs" }}
- name: gcs-log-creds-volume
secret:
secretName: {{ ternary (printf "%s-gcs-log-creds" ( .Release.Name )) .Values.global.storage.storageSecretName (not ((.Values.global.storage).storageSecretName)) }}
{{- end }}
While in airbyte-workload-api-server
chart deployment file
volumes:
{{- if and (eq .Values.global.deploymentMode "oss") (eq (lower (default "" .Values.global.storage.type)) "gcs") }}
- name: gcs-log-creds-volume
secret:
secretName: {{ ternary (printf "%s-gcs-log-creds" ( .Release.Name )) (.Values.global.credVolumeOverride) (eq .Values.global.deploymentMode "oss") }}
{{- end }}
It sounds the helm ternary function should looks like the one in airbyte-server
to use the .Values.global.storage.storageSecretName
if it has been provided in the values.yaml
...
secretName: {{ ternary (printf "%s-gcs-log-creds" ( .Release.Name )) .Values.global.storage.storageSecretName (not ((.Values.global.storage).storageSecretName)) }}
...
Regarding the small fix it might represent I was wondering if it was worth sending a PR.
airbytehq/airbyte