aglinxinyuan opened a new issue, #4157:
URL: https://github.com/apache/texera/issues/4157

   
   ### Discussed in https://github.com/apache/texera/discussions/4149
   
   <div type='discussions-op-text'>
   
   <sup>Originally posted by **aglinxinyuan** January  8, 2026</sup>
   I would like to introduce the idea of supporting multiple runtime execution 
modes that users can choose from based on the requirements of their use case 
and the characteristics of their jobs.
   
   The current (and default) execution behavior of our engine is what we call 
pipelined, or STREAMING, execution mode. In this mode, each operator performs 
continuous, incremental processing as data flows through the pipeline.
   
   In addition, we plan to support a batch-style execution mode, referred to as 
BATCH execution mode. This mode executes jobs in a manner more reminiscent of 
traditional batch processing. We intend to enable this mode via a configuration 
flag.
   
   Our unified approach to stream and batch processing ensures that 
applications executed over bounded inputs will produce the same final results 
regardless of the selected execution mode. Enabling BATCH execution allows the 
engine to apply additional optimizations that are only possible when operators 
know that their inputs are bounded. 
   
   Below, I will provide an example illustrating the differences between these 
execution modes:
   
   
![Streaming](https://github.com/user-attachments/assets/a0348591-0273-45b2-88d1-c31e37862ad4)
   
![Batch](https://github.com/user-attachments/assets/a33904ab-e407-4081-b811-3084c3288749)</div>


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]

Reply via email to