Sets the buffering timeout for data produced by this operation. The timeout defines how long
data may linger ina partially full buffer before being sent over the network.
Lower timeouts lead to lower tail latencies, but may affect throughput. Timeouts of 1 ms
still sustain high throughput, even for jobs with high parallelism.
A value of ‘-1’ means that the default buffer timeout should be used. A value of ‘0’
indicates that no buffering should happen, and all records/events should be immediately sent
through the network, without additional buffering.
timeout_millis – The maximum time between two output flushes.
The operator with buffer timeout set.