Github user HeartSaVioR commented on a diff in the pull request: https://github.com/apache/storm/pull/2241#discussion_r158007018 --- Diff: conf/defaults.yaml --- @@ -253,11 +278,17 @@ topology.trident.batch.emit.interval.millis: 500 topology.testing.always.try.serialize: false topology.classpath: null topology.environment: null -topology.bolts.outgoing.overflow.buffer.enable: false -topology.disruptor.wait.timeout.millis: 1000 -topology.disruptor.batch.size: 100 -topology.disruptor.batch.timeout.millis: 1 -topology.disable.loadaware.messaging: false + +topology.transfer.buffer.size: 1000 # size of recv queue for transfer worker thread +topology.transfer.batch.size: 1 # can be no larger than half of `topology.transfer.buffer.size` + +topology.executor.receive.buffer.size: 32768 # size of recv queue for spouts & bolts. Will be internally rounded up to next power of 2 (if not already a power of 2) +topology.producer.batch.size: 1 # can be no larger than half of `topology.executor.receive.buffer.size` + +topology.batch.flush.interval.millis: 1 # Flush tuples are disabled if this is set to 0 or if (topology.producer.batch.size=1 and topology.transfer.batch.size=1). +topology.spout.recvq.skips: 3 # Check recvQ once every N invocations of Spout's nextTuple() [when ACKs disabled] + +topology.disable.loadaware.messaging: false # load aware messaging can degrade throughput --- End diff -- May be better to describe the cases when we recommend using load aware messaging, or disable load aware messaging. The comment may mislead users to consider this option as always better to disable.
---