[ https://issues.apache.org/jira/browse/EAGLE-438?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15414313#comment-15414313 ]
Peter Kim commented on EAGLE-438: --------------------------------- Yes, in the config.json file that the script uses, the inputs will be an array of components as such "inputs": [ { "component": "namenode", "host": "127.0.0.1", "port": "50070", "https": false }, { "component": "resourcemanager", "host": "127.0.0.1", "port": "8088", "https": false }, { "component": "datanode", "host": "127.0.0.1", "port": "50075", "https": false } ] For your other question, I could also make it so that different components can send it to different kafka topics. So make the kafka output area in the json file into an array > Multiple Inputs for Hadoop JMX Collector Python Script > ------------------------------------------------------ > > Key: EAGLE-438 > URL: https://issues.apache.org/jira/browse/EAGLE-438 > Project: Eagle > Issue Type: New Feature > Reporter: Peter Kim > Priority: Trivial > Labels: features > > It would be very useful for the hadoop jmx collector to be able to collect > jmx metrics for multiple components at once. So here, I wish to extend the > default python script that collects Hadoop JMX Metrics to support multiple > inputs as oppose to only one. So, with this, one can collect all jmx metrics > for multiple components such as namenode, datanode, resource manager, > hmaster, etc. all at once. -- This message was sent by Atlassian JIRA (v6.3.4#6332)