Horia
10/27/2025, 5:41 PMkapa.ai
10/27/2025, 5:41 PMvalues.yaml file (if using Helm/Kubernetes):  yaml JOB_MAIN_CONTAINER_CPU_REQUEST: <value> JOB_MAIN_CONTAINER_CPU_LIMIT: <value> JOB_MAIN_CONTAINER_MEMORY_REQUEST: <value> JOB_MAIN_CONTAINER_MEMORY_LIMIT: <value> These settings ensure that each sync job container can request and use more CPU and memory, which may help if resource limits are the bottleneck. For Kubernetes deployments, you can also adjust pod resource requests and limits in the values.yaml file under the jobs.resources section:  yaml global:   jobs:     resources:       limits:         cpu: 2         memory: 4096Mi       requests:         cpu: 2         memory: 4096Mi Make sure these values are set high enough to utilize your EC2's capacity, but not so high that pods can't be scheduled due to cluster constraints. If you are running Airbyte in Docker, set the environment variables accordingly when starting the containers.   [Configuring Connector Resources](https://docs.airbyte.com/platform/operator-guides/configuring-connector-resources)  ### 2. Investigate Bottlenecks  - Low CPU Usage: If CPU usage is low and jobs are still hanging, the bottleneck may not be resource-related. It could be due to network latency, source/destination connector limitations, or issues with the underlying data (e.g., large documents, slow queries, or rate limits). - Logs and Monitoring: Check Airbyte logs for errors or warnings during the stalled periods. This can help identify if the job is waiting on the source, destination, or experiencing internal errors. - Connector-Specific Issues: Some connectors (like MongoDB) may have their own performance characteristics or limitations. For example, the source worker reads up to 10,000 records in memory, which can cause issues if records are large. If you have large documents, you may need to increase memory allocation for the source worker specifically [What To Scale](https://docs.airbyte.com/platform/operator-guides/scaling-airbyte#what-to-scale).  ### 3. Scaling and Concurrency  - Concurrent Syncs: Each sync job requires multiple pods/containers. If you run many syncs at once, ensure your cluster can handle the total number of pods. The rule of thumb is to allow for 2 x <number-of-possible-concurrent-connections> pods at once. - Worker Limits: You can control the number of concurrent syncs by setting MAX_SYNC_WORKERS and MAX_CHECK_WORKERS environment variables. If too many jobs run at once, some may be queued or starved of resources [Scaling Airbyte After Installation](https://docs.airbyte.com/platform/enterprise-setup/scaling-airbyte).  ### 4. Other Considerations  - Disk Space: While you have 1TB, ensure that disk I/O is not a bottleneck (e.g., EBS volume performance). - Network: If MongoDB or Redshift are remote, network latency or throughput could impact sync speed. - Connector Bugs: If only some syncs hang, there may be connector-specific bugs or issues with certain data sets.  ### Summary  - Explicitly set resource requests/limits for Airbyte job containers to ensure they can use your EC2's resources. - Monitor logs to identify where jobs are stalling. - Check for connector-specific limitations or bugs. - Adjust concurrency settings if running many syncs at once.  If after these steps the issue persists, consider sharing detailed logs from a stalled job for further diagnosis.  For more details, see the official docs on [Configuring Connector Resources](https://docs.airbyte.com/platform/operator-guides/configuring-connector-resources) and [Scaling Airbyte](https://docs.airbyte.com/platform/operator-guides/scaling-airbyte#what-to-scale).