Re: Monitoring the User Metrics for a long running Spark Job

2016-12-11 Thread Chawla,Sumit
Thanks a lot Sonal.. I will give it a try. Regards Sumit Chawla On Wed, Dec 7, 2016 at 10:45 PM, Sonal Goyal wrote: > You can try updating metrics.properties for the sink of your choice. In > our case, we add the following for getting application metrics in JSON >

Re: Monitoring the User Metrics for a long running Spark Job

2016-12-07 Thread Sonal Goyal
You can try updating metrics.properties for the sink of your choice. In our case, we add the following for getting application metrics in JSON format using http *.sink.reifier.class= org.apache.spark.metrics.sink.MetricsServlet Here, we have defined the sink with name reifier and its class is

Re: Monitoring the User Metrics for a long running Spark Job

2016-12-06 Thread Chawla,Sumit
Any pointers on this? Regards Sumit Chawla On Mon, Dec 5, 2016 at 8:30 PM, Chawla,Sumit wrote: > An example implementation i found is : https://github.com/groupon/ > spark-metrics > > Anyone has any experience using this? I am more interested in something > for

Re: Monitoring the User Metrics for a long running Spark Job

2016-12-05 Thread Chawla,Sumit
An example implementation i found is : https://github.com/groupon/spark-metrics Anyone has any experience using this? I am more interested in something for Pyspark specifically. The above link pointed to - https://github.com/apache/spark/blob/master/conf/metrics.properties.template. I need to

Re: Monitoring the User Metrics for a long running Spark Job

2016-12-05 Thread Miguel Morales
One thing I've done before is to install datadogs statsd agent on the nodes. Then you can emit metrics and stats to it and build dashboards on datadog. Sent from my iPhone > On Dec 5, 2016, at 8:17 PM, Chawla,Sumit wrote: > > Hi Manish > > I am specifically looking

Re: Monitoring the User Metrics for a long running Spark Job

2016-12-05 Thread Chawla,Sumit
Hi Manish I am specifically looking for something similar to following: https://ci.apache.org/projects/flink/flink-docs-release-1.1/apis/common/index.html#accumulators--counters. Flink has this concept of Accumulators, where user can keep its custom counters etc. While the application is

Re: Monitoring the User Metrics for a long running Spark Job

2016-12-05 Thread manish ranjan
http://spark.apache.org/docs/latest/monitoring.html You can even install tools like dstat , iostat , and iotop , *collectd* can provide fine-grained profiling on individual nodes. If

Monitoring the User Metrics for a long running Spark Job

2016-12-05 Thread Chawla,Sumit
Hi All I have a long running job which takes hours and hours to process data. How can i monitor the operational efficency of this job? I am interested in something like Storm\Flink style User metrics/aggregators, which i can monitor while my job is running. Using these metrics i want to