Re: Unable to populate spark metrics using custom metrics API

2023-07-13 Thread Surya Soma
Gentle reminder on this. On Sat, Jul 8, 2023 at 7:59 PM Surya Soma wrote: > Hello, > > I am trying to publish custom metrics using Spark CustomMetric API as > supported since spark 3.2 https://github.com/apache/spark/pull/31476, > > > https://spark.apache.org/docs/3.2.

Unable to populate spark metrics using custom metrics API

2023-07-08 Thread Surya Soma
Hello, I am trying to publish custom metrics using Spark CustomMetric API as supported since spark 3.2 https://github.com/apache/spark/pull/31476, https://spark.apache.org/docs/3.2.0/api/java/org/apache/spark/sql/connector/metric/CustomMetric.html I have created a custom metric implementing

Unable to populate spark metrics using custom metrics API

2023-06-26 Thread Surya Soma
Hello, I am trying to publish custom metrics using Spark CustomMetric API as supported since spark 3.2 https://github.com/apache/spark/pull/31476, https://spark.apache.org/docs/3.2.0/api/java/org/apache/spark/sql/connector/metric/CustomMetric.html I have created a custom metric implementing

Custom metrics in py-spark 3

2022-04-14 Thread Harut Martirosyan
Hello. We’re successfully exporting technical metrics to prometheus using built-in capabilities of Spark 3, but we need to add custom business metrics as well using python. Seems like there’s no documentation for that. Thanks.

Re: Best way to emit custom metrics to Prometheus in spark structured streaming

2020-11-04 Thread meetwes
So I tried it again in standalone mode (spark-shell) and the df.observe() functionality works. I tried sum, count, conditional aggregations using 'when', etc and all of this works in spark-shell. But, with spark-on-k8s, cluster mode, only using lit() as the aggregation column works. No other

Re: Best way to emit custom metrics to Prometheus in spark structured streaming

2020-11-04 Thread meetwes
Hi, Thanks for the reply. I tried it out today but I am unable to get it to work in cluster mode. The aggregation result is always 0. It works fine in standalone however with spark shell but with spark on Kubernetes in cluster mode, it doesn't. -- Sent from:

Re: Best way to emit custom metrics to Prometheus in spark structured streaming

2020-11-02 Thread Jungtaek Lim
You can try out "Dataset.observe" added in Spark 3, which enables arbitrary metrics to be logged and exposed to streaming query listeners. On Tue, Nov 3, 2020 at 3:25 AM meetwes wrote: > Hi I am looking for the right approach to emit custom metrics for spark > structured stream

Best way to emit custom metrics to Prometheus in spark structured streaming

2020-11-02 Thread meetwes
Hi I am looking for the right approach to emit custom metrics for spark structured streaming job.*Actual Scenario:* I have an aggregated dataframe let's say with (id, key, value) columns. One of the kpis could be 'droppedRecords' and the corresponding value column has the number of dropped records

Custom Metrics Source -> Sink routing

2020-09-30 Thread Dávid Szakállas
ush model is a better fit. We created a dedicated DropWizard MetricsRegistry for these custom metrics and are using https://github.com/banzaicloud/spark-metrics <https://github.com/banzaicloud/spark-metrics> for pushing the metrics to the PGW. However pushing all the metrics to the ga

Exporting spark custom metrics via prometheus jmx exporter

2020-09-21 Thread adilerman
I'm new to spark. Trying to export spark (2.4.0) custom metrics in prometheus format. I found this guide https://argus-sec.com/monitoring-spark-prometheus/ Running on my local pc (windows) I am able to see my metrics in visualVM and in spark's json http://localhost:4040/metrics/json/ My issue

Custom Metrics

2020-06-18 Thread Bryan Jeffrey
Hello. We're using Spark 2.4.4. We have a custom metrics sink consuming the Spark-produced metrics (e.g. heap free, etc.). I am trying to determine a good mechanism to pass the Spark application name into the metrics sink. Current the application ID is included, but not the application name

Re: Emit Custom metrics in Spark Structured Streaming job

2018-07-10 Thread subramgr
Hi I am working on implementing my idea but here is how it goes: 1. Use this library https://github.com/groupon/spark-metrics 2. Have a cron job which periodically curl /metrics/json endpoint at driver and all other nodes 3. Parse the response and send the data through a telegraf agent

Re: Emit Custom metrics in Spark Structured Streaming job

2018-07-10 Thread chandan prakash
Hi Subramanian, Did you find any solution for this ? I am looking for something similar too. Regards, Chandan On Wed, Jun 27, 2018 at 9:47 AM subramgr wrote: > I am planning to send these metrics to our KairosDB. Let me know if there > are > any examples that I can take a look > > > > -- >

Re: Emit Custom metrics in Spark Structured Streaming job

2018-06-26 Thread subramgr
I am planning to send these metrics to our KairosDB. Let me know if there are any examples that I can take a look -- Sent from: http://apache-spark-user-list.1001560.n3.nabble.com/ - To unsubscribe e-mail:

Emit Custom metrics in Spark Structured Streaming job

2018-06-26 Thread subramgr
In our Spark Structured Streaming job we listen to Kafka and we filter out some messages which we feel are malformed. Currently we log that information using the LOGGER. Is there a way to emit some kind of metrics for each time such a malformed message is seen in Structured Streaming ? Thanks

Re: Custom metrics sink

2018-03-16 Thread Silvio Fiorito
/apache/spark/metrics/sink/ConsoleSink.scala#L28 From: Christopher Piggott <cpigg...@gmail.com> Date: Friday, March 16, 2018 at 4:09 PM To: "user@spark.apache.org" <user@spark.apache.org> Subject: Custom metrics sink Just for fun, i want to make a stupid program that mak

Re: Custom metrics sink

2018-03-16 Thread Felix Cheung
There is a proposal to expose them. See SPARK-14151 From: Christopher Piggott <cpigg...@gmail.com> Sent: Friday, March 16, 2018 1:09:38 PM To: user@spark.apache.org Subject: Custom metrics sink Just for fun, i want to make a stupid program that makes dif

Custom metrics sink

2018-03-16 Thread Christopher Piggott
Just for fun, i want to make a stupid program that makes different frequency chimes as each worker becomes active. That way you can 'hear' what the cluster is doing and how it's distributing work. I thought to do this I would make a custom Sink, but the Sink and everything else in

EMR Spark Custom Metrics

2016-04-21 Thread Mark Kelly
Hi, So i would like some custom metrics. The environment we use is AWS EMR 4.5.0 with spark 1.6.1 and Ganglia. the code snippit below shows how we register custom metrics (this worked in EMR 4.2.0 with spark 1.5.2) package org.apache.spark.metrics.source import com.codahale.metrics._ import

How to get custom metrics using Ganglia Sink?

2015-12-08 Thread SRK
Hi, How do I configure custom metrics using Ganglia Sink? Thanks! -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/How-to-get-custom-metrics-using-Ganglia-Sink-tp25645.html Sent from the Apache Spark User List mailing list archive at Nabble.com

Re: Registering custom metrics

2015-06-23 Thread Otis Gospodnetić
Hi, Not sure if this will fit your needs, but if you are trying to collect+chart some metrics specific to your app, yet want to correlate them with what's going on in Spark, maybe Spark's performance numbers, you may want to send your custom metrics to SPM, so they can be visualized/analyzed

Re: Registering custom metrics

2015-06-22 Thread Dmitry Goldenberg
.. On Mon, Jun 22, 2015 at 10:14 AM, Silvio Fiorito silvio.fior...@granturing.com wrote: Sorry, replied to Gerard’s question vs yours. See here: Yes, you have to implement your own custom Metrics Source using the Code Hale library. See here for some examples: https://github.com/apache

Re: Registering custom metrics

2015-06-22 Thread dgoldenberg
Hi Gerard, Have there been any responses? Any insights as to what you ended up doing to enable custom metrics? I'm thinking of implementing a custom metrics sink, not sure how doable that is yet... Thanks. -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com

Re: Custom Metrics Sink

2015-06-22 Thread dgoldenberg
Hi, I was wondering if there've been any responses to this? Thanks. -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/Custom-Metrics-Sink-tp10068p23425.html Sent from the Apache Spark User List mailing list archive at Nabble.com

Re: Registering custom metrics

2015-06-22 Thread Silvio Fiorito
Hi Gerard, Yes, you have to implement your own custom Metrics Source using the Code Hale library. See here for some examples: https://github.com/apache/spark/blob/master/core/src/main/scala/org/apache/spark/metrics/source/JvmSource.scala https://github.com/apache/spark/blob/master/core/src/main

Re: Registering custom metrics

2015-06-22 Thread Silvio Fiorito
Sorry, replied to Gerard’s question vs yours. See here: Yes, you have to implement your own custom Metrics Source using the Code Hale library. See here for some examples: https://github.com/apache/spark/blob/master/core/src/main/scala/org/apache/spark/metrics/source/JvmSource.scala https

Registering Custom metrics [Spark-Streaming-monitoring]

2015-05-28 Thread Snehal Nagmote
Hello All, I am using spark streaming 1.3 . I want to capture few custom metrics based on accumulators, I followed somewhat similar to this approach , val instrumentation = new SparkInstrumentation(example.metrics) * val numReqs = sc.accumulator(0L

Re: Implementing custom metrics under MLPipeline's BinaryClassificationEvaluator

2015-05-18 Thread Joseph Bradley
/BinaryClassificationEvaluator.scala#L72 . Is my understanding correct? Or there are more convenient way of implementing a metric in order to be used by ML pipeline? Thanks. Justin -- View this message in context: Implementing custom metrics under MLPipeline's

Re: Registering custom metrics

2015-01-08 Thread Gerard Maas
) bytes } .saveAsTextFile(text) Is there a way to achieve this with the MetricSystem? ᐧ On Mon, Jan 5, 2015 at 10:24 AM, Gerard Maas gerard.m...@gmail.com wrote: Hi, Yes, I managed to create a register custom metrics by creating an implementation

Re: Registering custom metrics

2015-01-08 Thread Enno Shioji
, Gerard Maas gerard.m...@gmail.com wrote: Hi, Yes, I managed to create a register custom metrics by creating an implementation of org.apache.spark.metrics.source.Source and registering it to the metrics subsystem. Source is [Spark] private, so you need to create it under a org.apache.spark

Registering custom metrics

2014-10-30 Thread Gerard Maas
vHi, I've been exploring the metrics exposed by Spark and I'm wondering whether there's a way to register job-specific metrics that could be exposed through the existing metrics system. Would there be an example somewhere? BTW, documentation about how the metrics work could be improved. I

Custom Metrics Sink

2014-07-17 Thread jjaffe
What is the preferred way of adding a custom metrics sink to Spark? I noticed that the Sink Trait has been private since April, so I cannot simply extend Sink in an outside package, but I would like to avoid having to create a custom build of Spark. Is this possible? -- View this message