aglinxinyuan opened a new issue, #4157: URL: https://github.com/apache/texera/issues/4157
### Discussed in https://github.com/apache/texera/discussions/4149 <div type='discussions-op-text'> <sup>Originally posted by **aglinxinyuan** January 8, 2026</sup> I would like to introduce the idea of supporting multiple runtime execution modes that users can choose from based on the requirements of their use case and the characteristics of their jobs. The current (and default) execution behavior of our engine is what we call pipelined, or STREAMING, execution mode. In this mode, each operator performs continuous, incremental processing as data flows through the pipeline. In addition, we plan to support a batch-style execution mode, referred to as BATCH execution mode. This mode executes jobs in a manner more reminiscent of traditional batch processing. We intend to enable this mode via a configuration flag. Our unified approach to stream and batch processing ensures that applications executed over bounded inputs will produce the same final results regardless of the selected execution mode. Enabling BATCH execution allows the engine to apply additional optimizations that are only possible when operators know that their inputs are bounded. Below, I will provide an example illustrating the differences between these execution modes:  </div> -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: [email protected] For queries about this service, please contact Infrastructure at: [email protected]
