Hi Aurélien, Spark has endpoints to expose the spark application metrics. These endpoints can be used as a rest API. You can read more about these here: https://spark.apache.org/docs/3.1.1/monitoring.html#rest-api
Additionally, If you want to build your own custom metrics, you can explore spark custom plugins. Using a custom plugin, you can track your own custom metrics and plug it into the spark metrics system. Please note plugins are supported on spark versions above 3.0. -- Thanks & Regards, Akshay Haryani From: Aurélien Mazoyer <aurel...@aepsilon.com> Date: Thursday, September 2, 2021 at 8:36 AM To: user@spark.apache.org <user@spark.apache.org> Subject: Get application metric from Spark job Hi community, I would like to collect information about the execution of a Spark job while it is running. Could I define some kind of application metrics (such as a counter that would be incremented in my code) that I could retrieve regularly while the job is running? Thank you for help, Aurelien