Hello, We have a very large kafka topic (it gets ...
# general
c
Hello, We have a very large kafka topic (it gets about 10-25 mil rows/second) and we would like to use real time ingestion in order to fill in some tables. For each table we have a custom decoder that knows how to extract the proper data from each message or skip the message. I'm curious how the ingestion work - will Pinot stream the data independently or can we have one time ingestion/apply each decoder for the ingested rows? Thanks a lot :)
m
Sounds like you are overloading one kafka topic to have events (also schema?) across different tables? The issue with this approach is that every consumer of each Pinot table will have to ingest all the 10/25M rows and throw most of them.
I didn’t understand your question, could you elaborate a bit?