Hi Raman,
The banzaicloud jar can also cover the JMX exports.
Thanks,
Alex
On Fri, Sep 13, 2019 at 8:46 AM raman gugnani
wrote:
> Hi Alex,
>
> Thanks will check this out.
>
> Can it be done directly as spark also exposes the metrics or JVM. In
> this my one doubt is how to assign fixed JMX
Hi Alex,
Thanks will check this out.
Can it be done directly as spark also exposes the metrics or JVM. In this
my one doubt is how to assign fixed JMX ports to driver and executors.
@Alex,
Is there any difference in fetching data via JMX or using banzaicloud jar.
On Fri, 13 Sep 2019 at
Hi,
We are starting to use https://github.com/banzaicloud/spark-metrics .
Keep in mind that their solution is for Spark for K8s, to make it work for
Spark on Yarn you have to copy the dependencies of the spark-metrics into
Spark Jars folders on all the Spark machines (took me a while to figure).
Hi Team,
I am new to spark. I am using spark on hortonworks dataplatform with amazon
EC2 machines. I am running spark in cluster mode with yarn.
I need to monitor individual JVMs and other Spark metrics with *prometheus*.
Can anyone suggest the solution to do the same.
--
Raman Gugnani