Related, but not an identical question, is it possible to create a table from an external system, without specifying the schema at all?
Use case: I have a kafka topic with CDC data I’d like to store in my archive, and the underlying tables might change without Flink knowing about it (there are databases not in our control where we do have access to a CDC stream). I’d like the archive to contain all data, including of columns that were not there when the Flink job was first started.
Imagine the flow DB -> Kafka -> Flink -> Kafka -> Flink -> …
We’re currently using kafka connect to take a kafka topic and dump it into our archive, but this only really works for the first layer. @Anthony Daegele (Anthony) 1s soon as Flink is involved I would lose data (columns) until we adapt the schema and backfill the data.
Is there a way around this?