Hello, This may be a stupid question but is Arrow used for or designed with streaming processing use-cases in mind, where data is non-stationary. I.e: Flink stream processing jobs?
Particularly, is it possible from a given event source (say Kafka) to efficiently generate incremental record batches for stream processing? Suppose there is a data source that continuously generates messages with 100+ fields. You want to compute grouped aggregations (sums, averages, count distinct, etc...) over a select few of those fields, say 5 fields at most used for all queries. Is this a valid use-case for Arrow? What if time is important and some windowing technique has to be applied? Thank you very much for your time! Have a good day.