Precisely. Found a JIRA in this regard : SPARK-10610 <https://issues.apache.org/jira/browse/SPARK-10610>
On Wed, Mar 2, 2016 at 3:36 AM, Reynold Xin <r...@databricks.com> wrote: > Is the suggestion just to use a different config (and maybe fallback to > appid) in order to publish metrics? Seems reasonable. > > > On Tue, Mar 1, 2016 at 8:17 AM, Karan Kumar <karankumar1...@gmail.com> > wrote: > >> +dev mailing list >> >> Time series analysis on metrics becomes quite useful when running spark >> jobs using a workflow manager like oozie. >> >> Would love to take this up if the community thinks its worthwhile. >> >> On Tue, Feb 23, 2016 at 2:59 PM, Karan Kumar <karankumar1...@gmail.com> >> wrote: >> >>> HI >>> >>> Spark at the moment uses application ID to report metrics. I was >>> thinking that if we can create an option to export metrics on a >>> user-controlled key. This will allow us to do time series analysis on >>> counters by dumping these counters in a DB such as graphite. >>> >>> One of the approaches I had in mind was allowing a user to set a >>> property via the spark client. If that property is set, use the property >>> value to report metrics else use the current implementation >>> <https://github.com/apache/spark/blob/master/core/src/main/scala/org/apache/spark/metrics/MetricsSystem.scala>of >>> reporting metrics on appid. >>> >>> Thoughts? >>> >>> -- >>> Thanks >>> Karan >>> >> >> >> >> -- >> Thanks >> Karan >> > > -- Thanks Karan