Hi everyone! I'm new here using Airbyte and I don'...
# ask-community-for-troubleshooting
d
Hi everyone! I'm new here using Airbyte and I don't know how Airbyte pulls data from source to destination using Workers and Jobs. I've read these docs but I still don't understand how it works. Could someone give more details about it? • https://docs.airbyte.io/understanding-airbyte/high-level-viewhttps://docs.airbyte.io/understanding-airbyte/airbyte-specification For some reasons, I had to deploy Airbyte in an on-premise server and I want to move data between Snowflake and a Data Lake, both services are in the same region of AWS. I need to make sure how data is moving around in order to avoid data transfer costs between cloud and on-premise (egress). If Airbyte needs to download data to local and then push it to the cloud again, unfortunately this solution won't fit me. Thank you for having me here and your support! :)
a
yes, airbyte would pull data from the source to where you deployed airbyte, and then push to the destination. It's in a k8s context, but there is a diagram on this page that helped me understand: https://airbyte.com/blog/scaling-data-pipelines-kubernetes
d
Great, thank you! Just what I need to know. Then I guess I'll need to also deploy Airbyte in AWS to avoid moving data between cloud and local.