Kallin Nagelberg
10/16/2025, 7:10 PMpostgres source and snowflake destination.
Trying to do an initial sync (incremental | append + deduped). on a fairly narrow table with ~1mil rows. I've tried restarting a few times, and airbyte keeps getting 'stuck' part-way through the sync. I'm using a fairly beefy 16core 64gb ram ec2 instance, and it looks healthy. The last couple lines in the sync logs are like:
2025-10-15 17:53:04 source INFO main i.a.i.s.p.c.CtidStateManager(generateStateMessageAtCheckpoint):79 Emitting ctid state for stream public_accounts, state is io.airbyte.integrations.source.postgres.internal.models.CtidStatus@4a9486c0[version=2,stateType=ctid,ctid=(5996,66),incrementalState={"state_type":"cursor_based","version":2,"stream_name":"accounts","stream_namespace":"public","cursor_field":["id"],"cursor":"1057496","cursor_record_count":1},relationFilenode=54853845,additionalProperties={}]
2025-10-15 17:53:04 source INFO main i.a.c.i.s.r.s.SourceStateIterator(computeNext):50 sending state message, with count per stream: {public_accounts=10000}
2025-10-15 17:53:05 replication-orchestrator INFO Records read: 150000 (265 MB)
2025-10-15 17:53:07 replication-orchestrator INFO Records read: 155000 (274 MB)
and then nothing. There is no query running on postgres, nor on snowflake, so I imagine something is getting stalled out within airbyte. I wonder if it's possible there's something in the data that's causing it to fail? I can't find any additional logs, so will look into somehow setting up airbyte to produce more logs, but am somewhat out of ideas at the moment. This is my first experience/sync with airbyte, so feeling a bit disheartened at this point 💔kapa.ai
10/16/2025, 7:11 PMdocker logs <container_id> to get more details. Next Steps: - Upgrade both your Postgres source and Snowflake destination connectors to the latest versions. - Try increasing the buffer parameter in the Snowflake destination to 50. - Check disk space on your Airbyte workspace. - Try syncing a smaller subset of your data to isolate if a specific record is causing the issue. - Review full logs for any hidden errors or clues. - If possible, try syncing to a different destination (e.g., S3 or Redshift) to see if the issue is specific to Snowflake. If after these steps the issue persists, please consider sharing the full sync logs (with sensitive info redacted) on the Airbyte forum or GitHub for more targeted help. Let me know if you need guidance on any of these steps!Camille Lemaitre
10/17/2025, 8:07 AMKallin Nagelberg
10/17/2025, 1:20 PMKallin Nagelberg
10/17/2025, 5:25 PM# This controls the actual sync job pod containers
workloads:
resources:
# Set higher defaults for replication workloads
replication:
cpu:
request: "1"
limit: "4"
memory:
request: "8Gi"
limit: "16Gi"
it would have been nice to have gotten some message within airbyte that a worker was killed due to memory. I was using abctl, and would assume that I don't have to dive into K8s to see failures there. In case anyone from the product team is watching, this almost turned me off Airbyte as my first experience.