-
Notifications
You must be signed in to change notification settings - Fork 1.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Doc] How long till data is ready to be consumed at speed? #15036
Comments
btw: at operation of Pinot this kind of metric of "e2e freshness" would also have relevance in many use cases and would allow to
there is a "well aged" issue for this #4007 |
@Jackie-Jiang do you have any first example numbers on this / a source to share |
I'm not sure if I completely get the question, but I can answer from the perspective of how Pinot handles streaming data. Unlike a lot of other databases that ingest streaming data as mini batches (where the delay happens), Pinot directly writes the data into index row-by-row and the data immediately becomes queryable. The delay of streaming data arriving Pinot to it becoming queryable is usually below millisecond (Pinot can easily ingest thousands of messages per second). If you count end-to-end time from data produced to streaming system (e.g. Kafka) to becoming queryable in Pinot, the delay is usually a few seconds, and majority of the delay is from streaming system processing then delivering the messages to Pinot. |
Many thanks for this inside! Would be really interesting to have some real end2end benchmarks of durations |
Pinot can deliver query results with stunning speed /low latency which is described on many places
e.g. very nicely at startree s blog https://startree.ai/resources/what-makes-apache-pinot-fast-chapter-ii
In contrast its hard to find any numbers/examples on How long does it take to have the data ready to be consumed at this speed?
How long does it take from data ingest through the layers of Pinot including updating different index etc.
Would be handy to have some infos on this in the doc or blog or as a first step directly in this issue.
The text was updated successfully, but these errors were encountered: