Re: Get application metric from Spark job

2021-09-09 Thread Aurélien Mazoyer
Hi!

Again, thanks a lot for advice: I'll have a look!

Best,

Aurelien

Le mar. 7 sept. 2021 à 20:36, Haryani, Akshay  a
écrit :

> For custom metrics, you can take a look at Groupon’s spar metrics:
> https://github.com/groupon/spark-metrics
>
>
>
> It is supported on spark 2.x. Alternatively, you can create a custom
> source (extending source trait), enable the sink and register the custom
> source to get the metrics. Some useful links for this approach:
>
> https://gist.github.com/ibuenros/9b94736c2bad2f4b8e23
>
> https://kb.databricks.com/metrics/spark-metrics.html
>
>
> http://mail-archives.us.apache.org/mod_mbox/spark-user/201501.mbox/%3CCAE50=dq+6tdx9VNVM3ctBMWPLDPbUAacO3aN3L8x38zg=xb...@mail.gmail.com%3E
>
>
>
> I hope these help.
>
> --
>
> Thanks & Regards,
>
> Akshay Haryani
>
>
>
> *From: *Aurélien Mazoyer 
> *Date: *Monday, September 6, 2021 at 5:47 AM
> *To: *Haryani, Akshay 
> *Cc: *user@spark.apache.org 
> *Subject: *Re: Get application metric from Spark job
>
> Hi Akshay,
>
>
>
> Thank you for your reply. Sounds like a good idea, but I unfortunately
> have a 2.6 cluster. Do you know if there would be another solution that
> would run on 2.6 or if I have no other choice than migrating to 3?
>
>
>
> Regards,
>
>
>
> Aurélien
>
>
>
> Le jeu. 2 sept. 2021 à 20:12, Haryani, Akshay  a
> écrit :
>
> Hi Aurélien,
>
>
>
> Spark has endpoints to expose the spark application metrics. These
> endpoints can be used as a rest API. You can read more about these here:
> https://spark.apache.org/docs/3.1.1/monitoring.html#rest-api
>
>
>
> Additionally,
>
> If you want to build your own custom metrics, you can explore spark custom
> plugins. Using a custom plugin, you can track your own custom metrics and
> plug it into the spark metrics system. Please note plugins are supported
> on spark versions above 3.0.
>
>
>
>
>
> --
>
> Thanks & Regards,
>
> Akshay Haryani
>
>
>
> *From: *Aurélien Mazoyer 
> *Date: *Thursday, September 2, 2021 at 8:36 AM
> *To: *user@spark.apache.org 
> *Subject: *Get application metric from Spark job
>
> Hi community,
>
>
>
> I would like to collect information about the execution of a Spark job
> while it is running. Could I define some kind of application metrics (such
> as a counter that would be incremented in my code) that I could retrieve
> regularly while the job is running?
>
>
> Thank you for help,
>
>
>
> Aurelien
>
>


Re: Get application metric from Spark job

2021-09-07 Thread Haryani, Akshay
For custom metrics, you can take a look at Groupon’s spar metrics: 
https://github.com/groupon/spark-metrics

It is supported on spark 2.x. Alternatively, you can create a custom source 
(extending source trait), enable the sink and register the custom source to get 
the metrics. Some useful links for this approach:
https://gist.github.com/ibuenros/9b94736c2bad2f4b8e23
https://kb.databricks.com/metrics/spark-metrics.html
http://mail-archives.us.apache.org/mod_mbox/spark-user/201501.mbox/%3CCAE50=dq+6tdx9VNVM3ctBMWPLDPbUAacO3aN3L8x38zg=xb...@mail.gmail.com%3E

I hope these help.
--
Thanks & Regards,
Akshay Haryani

From: Aurélien Mazoyer 
Date: Monday, September 6, 2021 at 5:47 AM
To: Haryani, Akshay 
Cc: user@spark.apache.org 
Subject: Re: Get application metric from Spark job
Hi Akshay,

Thank you for your reply. Sounds like a good idea, but I unfortunately have a 
2.6 cluster. Do you know if there would be another solution that would run on 
2.6 or if I have no other choice than migrating to 3?

Regards,

Aurélien

Le jeu. 2 sept. 2021 à 20:12, Haryani, Akshay 
mailto:akshay.hary...@hpe.com>> a écrit :
Hi Aurélien,

Spark has endpoints to expose the spark application metrics. These endpoints 
can be used as a rest API. You can read more about these here: 
https://spark.apache.org/docs/3.1.1/monitoring.html#rest-api<https://spark.apache.org/docs/3.1.1/monitoring.html#rest-api>

Additionally,
If you want to build your own custom metrics, you can explore spark custom 
plugins. Using a custom plugin, you can track your own custom metrics and plug 
it into the spark metrics system. Please note plugins are supported on spark 
versions above 3.0.


--
Thanks & Regards,
Akshay Haryani

From: Aurélien Mazoyer mailto:aurel...@aepsilon.com>>
Date: Thursday, September 2, 2021 at 8:36 AM
To: user@spark.apache.org<mailto:user@spark.apache.org> 
mailto:user@spark.apache.org>>
Subject: Get application metric from Spark job
Hi community,

I would like to collect information about the execution of a Spark job while it 
is running. Could I define some kind of application metrics (such as a counter 
that would be incremented in my code) that I could retrieve regularly while the 
job is running?

Thank you for help,

Aurelien


Re: Get application metric from Spark job

2021-09-06 Thread Aurélien Mazoyer
Hi Akshay,

Thank you for your reply. Sounds like a good idea, but I unfortunately have
a 2.6 cluster. Do you know if there would be another solution that would
run on 2.6 or if I have no other choice than migrating to 3?

Regards,

Aurélien

Le jeu. 2 sept. 2021 à 20:12, Haryani, Akshay  a
écrit :

> Hi Aurélien,
>
>
>
> Spark has endpoints to expose the spark application metrics. These
> endpoints can be used as a rest API. You can read more about these here:
> https://spark.apache.org/docs/3.1.1/monitoring.html#rest-api
>
>
>
> Additionally,
>
> If you want to build your own custom metrics, you can explore spark custom
> plugins. Using a custom plugin, you can track your own custom metrics and
> plug it into the spark metrics system. Please note plugins are supported
> on spark versions above 3.0.
>
>
>
>
>
> --
>
> Thanks & Regards,
>
> Akshay Haryani
>
>
>
> *From: *Aurélien Mazoyer 
> *Date: *Thursday, September 2, 2021 at 8:36 AM
> *To: *user@spark.apache.org 
> *Subject: *Get application metric from Spark job
>
> Hi community,
>
>
>
> I would like to collect information about the execution of a Spark job
> while it is running. Could I define some kind of application metrics (such
> as a counter that would be incremented in my code) that I could retrieve
> regularly while the job is running?
>
>
> Thank you for help,
>
>
>
> Aurelien
>


Re: Get application metric from Spark job

2021-09-02 Thread Haryani, Akshay
Hi Aurélien,

Spark has endpoints to expose the spark application metrics. These endpoints 
can be used as a rest API. You can read more about these here: 
https://spark.apache.org/docs/3.1.1/monitoring.html#rest-api

Additionally,
If you want to build your own custom metrics, you can explore spark custom 
plugins. Using a custom plugin, you can track your own custom metrics and plug 
it into the spark metrics system. Please note plugins are supported on spark 
versions above 3.0.


--
Thanks & Regards,
Akshay Haryani

From: Aurélien Mazoyer 
Date: Thursday, September 2, 2021 at 8:36 AM
To: user@spark.apache.org 
Subject: Get application metric from Spark job
Hi community,

I would like to collect information about the execution of a Spark job while it 
is running. Could I define some kind of application metrics (such as a counter 
that would be incremented in my code) that I could retrieve regularly while the 
job is running?

Thank you for help,

Aurelien


Get application metric from Spark job

2021-09-02 Thread Aurélien Mazoyer
Hi community,

I would like to collect information about the execution of a Spark job
while it is running. Could I define some kind of application metrics (such
as a counter that would be incremented in my code) that I could retrieve
regularly while the job is running?

Thank you for help,

Aurelien