> On 9 Sep 2016, at 13:20, Han JU <ju.han.fe...@gmail.com> wrote: > > Hi, > > I'd like to know if there's a possibility to get spark's metrics from code. > For example > > val sc = new SparkContext(conf) > val result = myJob(sc, ...) > result.save(...) > > val gauge = MetricSystem.getGauge("org.apahce.spark....") > println(gauge.getValue) // or send to to internal aggregation service > > I'm aware that there's a configuration for sending metrics to several kinds > of sinks but I'm more interested in a per job basis style and we use a custom > log/metric aggregation service for building dashboards. >
It's all coda hale metrics; should be retrievable somehow, for a loose definition of "somehow" I'd be interested in knowing what you come up with here. --------------------------------------------------------------------- To unsubscribe e-mail: user-unsubscr...@spark.apache.org