So, I just created a table with >40k rows, but with daily segments, 318 segments in total - not good, I want to rollup to monthly segments later -, and defined a JSON index for my main columns which contain dynamic data (data that just can't be defined as static columns). Even trying to brutalize this thing by querying all the data with a limit that surpasses the amount of rows I still get ~600ms queries! Geez, this thing is fast! 🙂
m
Mayank
05/06/2022, 3:13 PM
Yes, it it fast 🙂. In your case, the data size seems small as well.
d
Diogo Baeder
05/06/2022, 3:14 PM
It's quite small, yes. 1 year of data, ~215 MB total size. It could easily fit a month of data for each segment - for larger regions of data for us this will be a good size.
p
Prashant Pandey
05/10/2022, 3:17 PM
Man we had like 19k segments in Pinot once 😄
fistbump 1
Prashant Pandey
05/10/2022, 3:17 PM
1G segments no less
d
Diogo Baeder
05/10/2022, 3:54 PM
Wow, that's quite a lot 😄
Did you end up using some partitioning for pruning segments?
p
Prashant Pandey
05/10/2022, 4:28 PM
We just used the default time based pruning on a column.