I am not sure what column/properties you are referring to.  But the event log in Spark deals with application level "events', not JVM-level metrics.  To retrieve the JVM metrics, you need to use the REST API provided in Spark.  Please see https://spark.apache.org/docs/latest/monitoring.html for details..

On 3/9/22 10:21 AM, Prasad Bhalerao wrote:
Hi,

I am trying to calculate CPU utilization of an Executor(JVM level CPU usage) using Event log. Can someone please help me with this?

1) Which column/properties to select
2) the correct formula to derive cpu usage

Has anyone done anything similar to this?

We have many pipelines and those are using very huge EMR clusters. I am trying to find out the cpu utilization and memory utilization of the nodes. This will help me find out if the clusters are under utilized and reduce the nodes,

Is there a better way to get these stats without changing the code?


Thanks,
Prasad


---------------------------------------------------------------------
To unsubscribe e-mail: user-unsubscr...@spark.apache.org

Reply via email to