-
Notifications
You must be signed in to change notification settings - Fork 846
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Feature request: window buffer #2963
Comments
Hi @danthegoodman1 does this guide satisfy your requirements or does it fall short somewhere? https://docs.redpanda.com/redpanda-connect/configuration/windowed_processing/ |
I didn’t happen to see that, my mistake. But that solution seems to be fixed tumbling windows, when really what I want is sliding windows (emit on every new record), unless I’m understanding the section on creating windows wrong? |
you can do sliding too: https://docs.redpanda.com/redpanda-connect/components/buffers/system_window/ |
Seems that probably does it! |
Processing on a single record for AI inference limits rpcn to a really small number of simple use cases.
Being able to window records by time and/or length would unlock many orders of magnitude more usecases with AI inference. Combined with #2962 it could be used for custom inference.
For example, you could have a window of 30 records/30 seconds and run continual sentiment analysis over a stream of data to see how it changes over time. Single record is not enough context to perform accurate sentiment analysis of a larger conversation/discussion.
Crucially, you need something like a
group_by
option that would allow the windowing to be grouped up by some JSON path so that you could do things per-discussion, per-tenant, etc. Otherwise all context would be merged together, which would not be very useful when contexts are dynamic (e.g. new discussions being created and abandoned).Can TTL the window after some time so it's not there forever consuming memory. No need to store bc it can be easily recovered on reboot if the data still exists.
The text was updated successfully, but these errors were encountered: