Github user roshannaik commented on the issue:
https://github.com/apache/storm/pull/2241
Some points covering prev comments by @HeartSaVioR and @revans2
**Throughput limiting:** That only makes sense if you are measuring
Throughput vs CPU/other resource usage. Latency measurements do not need it.
And its a sin if you are doing that when trying to measure throughput.
**TVL topology:**
- Given its rate limiting nature, it definitely does not have the right
name. Its employment of very high threads counts and rate limiting spouts
appear to be tuned to work within the limitations of the current msging system
and target the old sweetspot. Deserves a question. Harsha's measurements
(which are more sensible in terms of executor counts), shows that the current
msging was brought down to its knees very quickly once the rate limiting went
away.
@revans2
The drop you are seeing with the increased in splitter counts is indicative
of the increased CPU contention going on even when not enough data flowing
through an executor (the issue you initially brought up... of high CPU usage
for idle topos). The old system, executor seems to be spending more time
sleeping when there is insufficient data flow and less CPU contention and
adding redundant/idle executors is not affecting it as much.So you can
throughput plateaus.
Lowering the CPU contention for idle mode is something i plan to address...
and i think have left some TODOs for myself in the code already for to keep me
honest.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at [email protected] or file a JIRA ticket
with INFRA.
---