laila ribke
11/02/2022, 1:44 PMHarshith (Airbyte)
11/02/2022, 2:22 PMlaila ribke
11/02/2022, 3:58 PMlaila ribke
11/02/2022, 3:59 PMlaila ribke
11/02/2022, 4:00 PMlaila ribke
11/03/2022, 9:01 AMHarshith (Airbyte)
11/03/2022, 10:58 AMlaila ribke
11/03/2022, 12:33 PMEdward Gao (Airbyte)
11/03/2022, 12:37 PMlaila ribke
11/04/2022, 8:48 AMEdward Gao (Airbyte)
11/04/2022, 2:26 PM*_raw
tables, which just has your data in JSON blobs; normalized data will expand that out into normalized tables (and also handles deduping, if you’re using the incremental dedup+history sync mode)laila ribke
11/04/2022, 2:56 PMlaila ribke
11/04/2022, 2:59 PMEdward Gao (Airbyte)
11/04/2022, 3:17 PMlaila ribke
11/07/2022, 9:40 AMEdward Gao (Airbyte)
11/07/2022, 4:17 PMlaila ribke
11/07/2022, 4:48 PMEdward Gao (Airbyte)
11/07/2022, 5:21 PMCOPY
it into redshift). And then you’d be able to choose basic normalization or a custom dbt transformationlaila ribke
11/07/2022, 5:26 PMEdward Gao (Airbyte)
11/07/2022, 5:34 PM_raw
tables, but not the normalized tables? Or did you mean that no data got transferred at all?)
generally this setup feels a little redundant - the s3 -> redshift sync will do a second copy into s3, and then do a COPY into redshift. It would be more efficient to do a single GAds -> redshift sync directly, unless I’m misunderstanding the setup.
Creating 4 tables is (currently) baked into how normalization works - if that’s problematic cost-wise, then you’d probably need to roll your own custom transformation unfortunately, which would pull data from the _raw
table and generate a normalized table. (but certainly this is useful feedback, I’ll pass it on to the rest of the team)laila ribke
11/07/2022, 6:16 PMlaila ribke
11/07/2022, 6:17 PMEdward Gao (Airbyte)
11/07/2022, 6:41 PMselect only data I really needinteresting! is this at a greater granularity than what the google ads source provides? I.e. you’re looking for more control than just “sync streams X, Y, Z”?
logthere’s an auth layer in front of this (which is good!) but means I can’t access it 🙂 can you download the logs and upload them here?
laila ribke
11/07/2022, 7:00 PMEdward Gao (Airbyte)
11/07/2022, 8:33 PM2022-11-07 18:47:08
- 'filetype': 'csv'
) I’m not super familiar with sourec-s3, but IIRC there’s a config option to have it read avro instead
and of course - it’s always useful to doublecheck that the files on S3 match whatever path is configured on sourec-s3, just in caselaila ribke
11/07/2022, 9:22 PMEdward Gao (Airbyte)
11/07/2022, 10:55 PMlaila ribke
11/08/2022, 8:27 AMlaila ribke
11/08/2022, 11:31 AMEdward Gao (Airbyte)
11/08/2022, 5:17 PMEdward Gao (Airbyte)
11/08/2022, 5:23 PMEdward Gao (Airbyte)
11/09/2022, 4:00 PMgaS3
stream (click the little arrow on the left) to confirm that it has the expected fields?