Hi there, a quick question regarding the S3 destination:
I am using an EC2 instance with 16 or 32 Gigs of RAM to pull in a data source which, in total, provides several 100 Gigs uncompressed.
I see in the logs what the total amount of data read is and when a buffer is flushed. I am using the default S3 destination settings for Parquet SNAPPY compression.
Yet, every single run, the sync fails with OoM errors.
Why would that happen if the buffer flushes every 200MB or so (default setting is 128MB even)?
Am I missing something obvious why memory is drowning besides flushes?