[ https://issues.apache.org/jira/browse/LOG4J2-2937?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17203734#comment-17203734 ]
Ralph Goers commented on LOG4J2-2937: ------------------------------------- I have no idea how splunk is implemented as it is a proprietary product, but since its primary purpose is as a logging query engine and logs are generally in time sequence I would imaging it has that capability built in. You can google Splun and look at its command reference for all the various things you can do like timechart and streamstats (both of which I use heavily) and see how powerful it is. Yes, Splunk polls once per minute. If it ever goes down (a rarity) you would have a gap. Yes, that is true about granularity. I have set splunk to poll once per minute. The overhead of that is miniscule. Yes, to try to do more would impact the application too much. Even products like New Relic or Datadog try to minimize the amount of work done in application threads. They have separate threads where they push data to their central servers. They may do more work on the threads to embellish the data. > Provide counters to measure log rate > ------------------------------------ > > Key: LOG4J2-2937 > URL: https://issues.apache.org/jira/browse/LOG4J2-2937 > Project: Log4j 2 > Issue Type: New Feature > Reporter: Dennys Fredericci > Priority: Major > Attachments: image-2020-09-28-21-10-13-850.png > > > As a Log4j API user will be really nice to have a way to get the number of > log calls for each level without any instrumentation or bytecode > manipulation, something native from log4j API. > Once this interface is implemented this can be exposed through JMX or used by > other libraries to send the log rate to monitoring systems such as Datadog, > NewRelic, Dynatrace, etc. :) -- This message was sent by Atlassian Jira (v8.3.4#803005)