Hi All, In my MySql ->Redshift (S3 staging) con...
# ask-community-for-troubleshooting
l
Hi All, In my MySql ->Redshift (S3 staging) connection I see that source emitted 10 million rows of clickout table. But in destination, I have only 400K. I´m using incremental refresh. This source was used in the past with other destinations. May that be the problem why I receive only 400K rows? For me, it´s a new connection, so it should fully refresh, and only next time increment. But it seems it thinks it has only to increment.
u
Hi laila, could you check that all records are present in the raw table? If they are, this could be a normalization issue and you'd need to see what is the commonality between all these records - a pattern in the schema.
l
Hi, I even created a new connection. The source emits 10M rows, but I receive in destination only 3M rows. It´s a MySql->S3->Redshift connection. I even did a full refresh Overwrite. Attaching the logs. What can it be?
n
If you look in the logs you'll see that all records emitted were committed. This means the issue is not in the connection/sync running. Could you try setting up a new MySQL source and try a sync on a small subset of the data to see if the error persists?
l
In the same sync, I synced 5 tables. On 4 of them I received all records. Only in this one I´m missing records. Might it be because of it´s size (10 million records)?
Hi.. I don´t understand. The record number emitted a committed is correct, but I don´t see them in my destination.. Source emitted the 10M rows.. May it be something with the S3 strategy?
n
Hello! Were you able to do a test sync with a subset of the data?
l
I don't have the possibility. But I checked that the source has no filters.. I'm attaching the S3 file and the sync logs. The rows count of the raw table and the normalized one is equal
Hi, just to let you know we solved the problem. In the destination, we set the S3 Filename pattern as {sync_id}. As it´s a large table, airbyte sends to S3 batches. And they were overwritten. We changed the Filename pattern to {sync_id}_{timestamp:millis} and S3 stored several files for that table.
n
Thank you so much for following up and I'm so glad to hear you've resolved it!