I have a aws glue zero-etl pointed at an s3 table ...
# ask-community-for-troubleshooting
b
I have a aws glue zero-etl pointed at an s3 table bucket. I want to use airbyte to sync an additional table that’s not yet supported by aws zero-etl. What’s the recommended way to push incemental updated data to an s3 table bucket? The s3 data lake destination using glue?
k
Yes, the recommended way to push incremental updated data to an S3 table bucket—especially if you want to integrate with AWS Glue—is to use the S3 Data Lake destination in Airbyte, configured with the Glue catalog. The S3 Data Lake destination is Airbyte’s official connector for writing Iceberg tables to S3, and it supports integration with AWS Glue as the catalog. This setup allows you to manage your tables and metadata in Glue, making them accessible to other AWS analytics services (like Athena, Redshift Spectrum, etc.) and compatible with your zero-ETL workflows. Key steps: 1. Set up S3 and Glue permissions: - Create an S3 bucket policy with both object-level and bucket-level permissions. - Update the policy to include the necessary Glue permissions (such as
glue:CreateTable
,
glue:GetTable
, etc.) for managing Iceberg tables via Glue. - If using Lake Formation, grant the required permissions for data location and database/table creation as needed. [See detailed policy example and steps here.](https://docs.airbyte.com/integrations/destinations/s3-data-lake#iceberg-catalog-setup-and-permissions) 2. Configure the S3 Data Lake destination in Airbyte: - Select the S3 Data Lake destination and choose Glue as the catalog. - Provide the warehouse location (e.g.,
s3://<bucket name>/path/within/bucket
). - Enter your AWS credentials or Role ARN as appropriate. 3. Set up your Airbyte connection: - Choose your source and the S3 Data Lake destination. - Configure the sync mode as Incremental - Append (or Deduped, if supported and needed). - Airbyte will then push only the new or updated records to your S3 bucket, and the Glue catalog will keep the metadata up to date. Why not use the standard S3 destination? The standard S3 destination writes files (CSV, Parquet, etc.) to S3, but does not manage table metadata in Glue or support Iceberg tables. For integration with Glue and modern data lakehouse features, the S3 Data Lake destination is preferred. Summary: Use the S3 Data Lake destination with the Glue catalog in Airbyte to push incremental updates to your S3 bucket and have them managed as Iceberg tables in Glue. This approach is designed for exactly your use case—augmenting your data lake with additional tables not yet supported by AWS zero-ETL, while maintaining compatibility with Glue and downstream analytics tools. For more details and setup instructions, refer to the [S3 Data Lake destination documentation](https://docs.airbyte.com/integrations/destinations/s3-data-lake#iceberg-catalog-setup-and-permissions).