[jira] [Updated] (HUDI-867) Graphite metrics are throwing IllegalArgumentException on continuous mode
[ https://issues.apache.org/jira/browse/HUDI-867?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Pratyaksh Sharma updated HUDI-867: -- Status: Closed (was: Patch Available) > Graphite metrics are throwing IllegalArgumentException on continuous mode > - > > Key: HUDI-867 > URL: https://issues.apache.org/jira/browse/HUDI-867 > Project: Apache Hudi > Issue Type: Bug > Components: DeltaStreamer >Reporter: João Esteves >Assignee: Pratyaksh Sharma >Priority: Major > Labels: bug-bash-0.6.0, pull-request-available > > Hello everyone, I am trying to extract Graphite metrics from Hudi using a > Spark Streaming process, but the method that sends metrics is throwing > java.lang.IllegalArgumentException after the first microbatch, like this: > {code:java} > 20/05/06 11:49:25 ERROR Metrics: Failed to send metrics: > java.lang.IllegalArgumentException: A metric named > kafka_hudi.finalize.duration already exists > at > org.apache.hudi.com.codahale.metrics.MetricRegistry.register(MetricRegistry.java:97) > at org.apache.hudi.metrics.Metrics.registerGauge(Metrics.java:83) > at > org.apache.hudi.metrics.HoodieMetrics.updateFinalizeWriteMetrics(HoodieMetrics.java:177) > at > org.apache.hudi.HoodieWriteClient.lambda$finalizeWrite$14(HoodieWriteClient.java:1233) > at org.apache.hudi.common.util.Option.ifPresent(Option.java:96) > at > org.apache.hudi.HoodieWriteClient.finalizeWrite(HoodieWriteClient.java:1231) > at org.apache.hudi.HoodieWriteClient.commit(HoodieWriteClient.java:497) > at org.apache.hudi.HoodieWriteClient.commit(HoodieWriteClient.java:479) > at org.apache.hudi.HoodieWriteClient.commit(HoodieWriteClient.java:470) > at > org.apache.hudi.HoodieSparkSqlWriter$.write(HoodieSparkSqlWriter.scala:152) > at > org.apache.hudi.HoodieStreamingSink$$anonfun$1$$anonfun$2.apply(HoodieStreamingSink.scala:51) > at > org.apache.hudi.HoodieStreamingSink$$anonfun$1$$anonfun$2.apply(HoodieStreamingSink.scala:51) > at scala.util.Try$.apply(Try.scala:192) > at > org.apache.hudi.HoodieStreamingSink$$anonfun$1.apply(HoodieStreamingSink.scala:50) > at > org.apache.hudi.HoodieStreamingSink$$anonfun$1.apply(HoodieStreamingSink.scala:50) > at > org.apache.hudi.HoodieStreamingSink.retry(HoodieStreamingSink.scala:114) > at > org.apache.hudi.HoodieStreamingSink.addBatch(HoodieStreamingSink.scala:49) > at > org.apache.spark.sql.execution.streaming.MicroBatchExecution$$anonfun$org$apache$spark$sql$execution$streaming$MicroBatchExecution$$runBatch$5$$anonfun$apply$17.apply(MicroBatchExecution.scala:537) > at > org.apache.spark.sql.execution.SQLExecution$$anonfun$withNewExecutionId$1.apply(SQLExecution.scala:84) > at > org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:165) > at > org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:74) > at > org.apache.spark.sql.execution.streaming.MicroBatchExecution$$anonfun$org$apache$spark$sql$execution$streaming$MicroBatchExecution$$runBatch$5.apply(MicroBatchExecution.scala:535) > at > org.apache.spark.sql.execution.streaming.ProgressReporter$class.reportTimeTaken(ProgressReporter.scala:351) > at > org.apache.spark.sql.execution.streaming.StreamExecution.reportTimeTaken(StreamExecution.scala:58) > at > org.apache.spark.sql.execution.streaming.MicroBatchExecution.org$apache$spark$sql$execution$streaming$MicroBatchExecution$$runBatch(MicroBatchExecution.scala:534) > at > org.apache.spark.sql.execution.streaming.MicroBatchExecution$$anonfun$runActivatedStream$1$$anonfun$apply$mcZ$sp$1.apply$mcV$sp(MicroBatchExecution.scala:198) > at > org.apache.spark.sql.execution.streaming.MicroBatchExecution$$anonfun$runActivatedStream$1$$anonfun$apply$mcZ$sp$1.apply(MicroBatchExecution.scala:166) > at > org.apache.spark.sql.execution.streaming.MicroBatchExecution$$anonfun$runActivatedStream$1$$anonfun$apply$mcZ$sp$1.apply(MicroBatchExecution.scala:166) > at > org.apache.spark.sql.execution.streaming.ProgressReporter$class.reportTimeTaken(ProgressReporter.scala:351) > at > org.apache.spark.sql.execution.streaming.StreamExecution.reportTimeTaken(StreamExecution.scala:58) > at > org.apache.spark.sql.execution.streaming.MicroBatchExecution$$anonfun$runActivatedStream$1.apply$mcZ$sp(MicroBatchExecution.scala:166) > at > org.apache.spark.sql.execution.streaming.ProcessingTimeExecutor.execute(TriggerExecutor.scala:56) > at > org.apache.spark.sql.execution.streaming.MicroBatchExecution.runActivatedStream(MicroBatchExecution.scala:160) > at >
[jira] [Updated] (HUDI-867) Graphite metrics are throwing IllegalArgumentException on continuous mode
[ https://issues.apache.org/jira/browse/HUDI-867?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] sivabalan narayanan updated HUDI-867: - Status: Patch Available (was: In Progress) > Graphite metrics are throwing IllegalArgumentException on continuous mode > - > > Key: HUDI-867 > URL: https://issues.apache.org/jira/browse/HUDI-867 > Project: Apache Hudi > Issue Type: Bug > Components: DeltaStreamer >Reporter: João Esteves >Assignee: Pratyaksh Sharma >Priority: Major > Labels: bug-bash-0.6.0, pull-request-available > > Hello everyone, I am trying to extract Graphite metrics from Hudi using a > Spark Streaming process, but the method that sends metrics is throwing > java.lang.IllegalArgumentException after the first microbatch, like this: > {code:java} > 20/05/06 11:49:25 ERROR Metrics: Failed to send metrics: > java.lang.IllegalArgumentException: A metric named > kafka_hudi.finalize.duration already exists > at > org.apache.hudi.com.codahale.metrics.MetricRegistry.register(MetricRegistry.java:97) > at org.apache.hudi.metrics.Metrics.registerGauge(Metrics.java:83) > at > org.apache.hudi.metrics.HoodieMetrics.updateFinalizeWriteMetrics(HoodieMetrics.java:177) > at > org.apache.hudi.HoodieWriteClient.lambda$finalizeWrite$14(HoodieWriteClient.java:1233) > at org.apache.hudi.common.util.Option.ifPresent(Option.java:96) > at > org.apache.hudi.HoodieWriteClient.finalizeWrite(HoodieWriteClient.java:1231) > at org.apache.hudi.HoodieWriteClient.commit(HoodieWriteClient.java:497) > at org.apache.hudi.HoodieWriteClient.commit(HoodieWriteClient.java:479) > at org.apache.hudi.HoodieWriteClient.commit(HoodieWriteClient.java:470) > at > org.apache.hudi.HoodieSparkSqlWriter$.write(HoodieSparkSqlWriter.scala:152) > at > org.apache.hudi.HoodieStreamingSink$$anonfun$1$$anonfun$2.apply(HoodieStreamingSink.scala:51) > at > org.apache.hudi.HoodieStreamingSink$$anonfun$1$$anonfun$2.apply(HoodieStreamingSink.scala:51) > at scala.util.Try$.apply(Try.scala:192) > at > org.apache.hudi.HoodieStreamingSink$$anonfun$1.apply(HoodieStreamingSink.scala:50) > at > org.apache.hudi.HoodieStreamingSink$$anonfun$1.apply(HoodieStreamingSink.scala:50) > at > org.apache.hudi.HoodieStreamingSink.retry(HoodieStreamingSink.scala:114) > at > org.apache.hudi.HoodieStreamingSink.addBatch(HoodieStreamingSink.scala:49) > at > org.apache.spark.sql.execution.streaming.MicroBatchExecution$$anonfun$org$apache$spark$sql$execution$streaming$MicroBatchExecution$$runBatch$5$$anonfun$apply$17.apply(MicroBatchExecution.scala:537) > at > org.apache.spark.sql.execution.SQLExecution$$anonfun$withNewExecutionId$1.apply(SQLExecution.scala:84) > at > org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:165) > at > org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:74) > at > org.apache.spark.sql.execution.streaming.MicroBatchExecution$$anonfun$org$apache$spark$sql$execution$streaming$MicroBatchExecution$$runBatch$5.apply(MicroBatchExecution.scala:535) > at > org.apache.spark.sql.execution.streaming.ProgressReporter$class.reportTimeTaken(ProgressReporter.scala:351) > at > org.apache.spark.sql.execution.streaming.StreamExecution.reportTimeTaken(StreamExecution.scala:58) > at > org.apache.spark.sql.execution.streaming.MicroBatchExecution.org$apache$spark$sql$execution$streaming$MicroBatchExecution$$runBatch(MicroBatchExecution.scala:534) > at > org.apache.spark.sql.execution.streaming.MicroBatchExecution$$anonfun$runActivatedStream$1$$anonfun$apply$mcZ$sp$1.apply$mcV$sp(MicroBatchExecution.scala:198) > at > org.apache.spark.sql.execution.streaming.MicroBatchExecution$$anonfun$runActivatedStream$1$$anonfun$apply$mcZ$sp$1.apply(MicroBatchExecution.scala:166) > at > org.apache.spark.sql.execution.streaming.MicroBatchExecution$$anonfun$runActivatedStream$1$$anonfun$apply$mcZ$sp$1.apply(MicroBatchExecution.scala:166) > at > org.apache.spark.sql.execution.streaming.ProgressReporter$class.reportTimeTaken(ProgressReporter.scala:351) > at > org.apache.spark.sql.execution.streaming.StreamExecution.reportTimeTaken(StreamExecution.scala:58) > at > org.apache.spark.sql.execution.streaming.MicroBatchExecution$$anonfun$runActivatedStream$1.apply$mcZ$sp(MicroBatchExecution.scala:166) > at > org.apache.spark.sql.execution.streaming.ProcessingTimeExecutor.execute(TriggerExecutor.scala:56) > at > org.apache.spark.sql.execution.streaming.MicroBatchExecution.runActivatedStream(MicroBatchExecution.scala:160) > at >
[jira] [Updated] (HUDI-867) Graphite metrics are throwing IllegalArgumentException on continuous mode
[ https://issues.apache.org/jira/browse/HUDI-867?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] sivabalan narayanan updated HUDI-867: - Status: In Progress (was: Open) > Graphite metrics are throwing IllegalArgumentException on continuous mode > - > > Key: HUDI-867 > URL: https://issues.apache.org/jira/browse/HUDI-867 > Project: Apache Hudi > Issue Type: Bug > Components: DeltaStreamer >Reporter: João Esteves >Assignee: Pratyaksh Sharma >Priority: Major > Labels: bug-bash-0.6.0, pull-request-available > > Hello everyone, I am trying to extract Graphite metrics from Hudi using a > Spark Streaming process, but the method that sends metrics is throwing > java.lang.IllegalArgumentException after the first microbatch, like this: > {code:java} > 20/05/06 11:49:25 ERROR Metrics: Failed to send metrics: > java.lang.IllegalArgumentException: A metric named > kafka_hudi.finalize.duration already exists > at > org.apache.hudi.com.codahale.metrics.MetricRegistry.register(MetricRegistry.java:97) > at org.apache.hudi.metrics.Metrics.registerGauge(Metrics.java:83) > at > org.apache.hudi.metrics.HoodieMetrics.updateFinalizeWriteMetrics(HoodieMetrics.java:177) > at > org.apache.hudi.HoodieWriteClient.lambda$finalizeWrite$14(HoodieWriteClient.java:1233) > at org.apache.hudi.common.util.Option.ifPresent(Option.java:96) > at > org.apache.hudi.HoodieWriteClient.finalizeWrite(HoodieWriteClient.java:1231) > at org.apache.hudi.HoodieWriteClient.commit(HoodieWriteClient.java:497) > at org.apache.hudi.HoodieWriteClient.commit(HoodieWriteClient.java:479) > at org.apache.hudi.HoodieWriteClient.commit(HoodieWriteClient.java:470) > at > org.apache.hudi.HoodieSparkSqlWriter$.write(HoodieSparkSqlWriter.scala:152) > at > org.apache.hudi.HoodieStreamingSink$$anonfun$1$$anonfun$2.apply(HoodieStreamingSink.scala:51) > at > org.apache.hudi.HoodieStreamingSink$$anonfun$1$$anonfun$2.apply(HoodieStreamingSink.scala:51) > at scala.util.Try$.apply(Try.scala:192) > at > org.apache.hudi.HoodieStreamingSink$$anonfun$1.apply(HoodieStreamingSink.scala:50) > at > org.apache.hudi.HoodieStreamingSink$$anonfun$1.apply(HoodieStreamingSink.scala:50) > at > org.apache.hudi.HoodieStreamingSink.retry(HoodieStreamingSink.scala:114) > at > org.apache.hudi.HoodieStreamingSink.addBatch(HoodieStreamingSink.scala:49) > at > org.apache.spark.sql.execution.streaming.MicroBatchExecution$$anonfun$org$apache$spark$sql$execution$streaming$MicroBatchExecution$$runBatch$5$$anonfun$apply$17.apply(MicroBatchExecution.scala:537) > at > org.apache.spark.sql.execution.SQLExecution$$anonfun$withNewExecutionId$1.apply(SQLExecution.scala:84) > at > org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:165) > at > org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:74) > at > org.apache.spark.sql.execution.streaming.MicroBatchExecution$$anonfun$org$apache$spark$sql$execution$streaming$MicroBatchExecution$$runBatch$5.apply(MicroBatchExecution.scala:535) > at > org.apache.spark.sql.execution.streaming.ProgressReporter$class.reportTimeTaken(ProgressReporter.scala:351) > at > org.apache.spark.sql.execution.streaming.StreamExecution.reportTimeTaken(StreamExecution.scala:58) > at > org.apache.spark.sql.execution.streaming.MicroBatchExecution.org$apache$spark$sql$execution$streaming$MicroBatchExecution$$runBatch(MicroBatchExecution.scala:534) > at > org.apache.spark.sql.execution.streaming.MicroBatchExecution$$anonfun$runActivatedStream$1$$anonfun$apply$mcZ$sp$1.apply$mcV$sp(MicroBatchExecution.scala:198) > at > org.apache.spark.sql.execution.streaming.MicroBatchExecution$$anonfun$runActivatedStream$1$$anonfun$apply$mcZ$sp$1.apply(MicroBatchExecution.scala:166) > at > org.apache.spark.sql.execution.streaming.MicroBatchExecution$$anonfun$runActivatedStream$1$$anonfun$apply$mcZ$sp$1.apply(MicroBatchExecution.scala:166) > at > org.apache.spark.sql.execution.streaming.ProgressReporter$class.reportTimeTaken(ProgressReporter.scala:351) > at > org.apache.spark.sql.execution.streaming.StreamExecution.reportTimeTaken(StreamExecution.scala:58) > at > org.apache.spark.sql.execution.streaming.MicroBatchExecution$$anonfun$runActivatedStream$1.apply$mcZ$sp(MicroBatchExecution.scala:166) > at > org.apache.spark.sql.execution.streaming.ProcessingTimeExecutor.execute(TriggerExecutor.scala:56) > at > org.apache.spark.sql.execution.streaming.MicroBatchExecution.runActivatedStream(MicroBatchExecution.scala:160) > at >
[jira] [Updated] (HUDI-867) Graphite metrics are throwing IllegalArgumentException on continuous mode
[ https://issues.apache.org/jira/browse/HUDI-867?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] ASF GitHub Bot updated HUDI-867: Labels: bug-bash-0.6.0 pull-request-available (was: bug-bash-0.6.0) > Graphite metrics are throwing IllegalArgumentException on continuous mode > - > > Key: HUDI-867 > URL: https://issues.apache.org/jira/browse/HUDI-867 > Project: Apache Hudi (incubating) > Issue Type: Bug > Components: DeltaStreamer >Reporter: João Esteves >Assignee: Pratyaksh Sharma >Priority: Major > Labels: bug-bash-0.6.0, pull-request-available > > Hello everyone, I am trying to extract Graphite metrics from Hudi using a > Spark Streaming process, but the method that sends metrics is throwing > java.lang.IllegalArgumentException after the first microbatch, like this: > {code:java} > 20/05/06 11:49:25 ERROR Metrics: Failed to send metrics: > java.lang.IllegalArgumentException: A metric named > kafka_hudi.finalize.duration already exists > at > org.apache.hudi.com.codahale.metrics.MetricRegistry.register(MetricRegistry.java:97) > at org.apache.hudi.metrics.Metrics.registerGauge(Metrics.java:83) > at > org.apache.hudi.metrics.HoodieMetrics.updateFinalizeWriteMetrics(HoodieMetrics.java:177) > at > org.apache.hudi.HoodieWriteClient.lambda$finalizeWrite$14(HoodieWriteClient.java:1233) > at org.apache.hudi.common.util.Option.ifPresent(Option.java:96) > at > org.apache.hudi.HoodieWriteClient.finalizeWrite(HoodieWriteClient.java:1231) > at org.apache.hudi.HoodieWriteClient.commit(HoodieWriteClient.java:497) > at org.apache.hudi.HoodieWriteClient.commit(HoodieWriteClient.java:479) > at org.apache.hudi.HoodieWriteClient.commit(HoodieWriteClient.java:470) > at > org.apache.hudi.HoodieSparkSqlWriter$.write(HoodieSparkSqlWriter.scala:152) > at > org.apache.hudi.HoodieStreamingSink$$anonfun$1$$anonfun$2.apply(HoodieStreamingSink.scala:51) > at > org.apache.hudi.HoodieStreamingSink$$anonfun$1$$anonfun$2.apply(HoodieStreamingSink.scala:51) > at scala.util.Try$.apply(Try.scala:192) > at > org.apache.hudi.HoodieStreamingSink$$anonfun$1.apply(HoodieStreamingSink.scala:50) > at > org.apache.hudi.HoodieStreamingSink$$anonfun$1.apply(HoodieStreamingSink.scala:50) > at > org.apache.hudi.HoodieStreamingSink.retry(HoodieStreamingSink.scala:114) > at > org.apache.hudi.HoodieStreamingSink.addBatch(HoodieStreamingSink.scala:49) > at > org.apache.spark.sql.execution.streaming.MicroBatchExecution$$anonfun$org$apache$spark$sql$execution$streaming$MicroBatchExecution$$runBatch$5$$anonfun$apply$17.apply(MicroBatchExecution.scala:537) > at > org.apache.spark.sql.execution.SQLExecution$$anonfun$withNewExecutionId$1.apply(SQLExecution.scala:84) > at > org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:165) > at > org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:74) > at > org.apache.spark.sql.execution.streaming.MicroBatchExecution$$anonfun$org$apache$spark$sql$execution$streaming$MicroBatchExecution$$runBatch$5.apply(MicroBatchExecution.scala:535) > at > org.apache.spark.sql.execution.streaming.ProgressReporter$class.reportTimeTaken(ProgressReporter.scala:351) > at > org.apache.spark.sql.execution.streaming.StreamExecution.reportTimeTaken(StreamExecution.scala:58) > at > org.apache.spark.sql.execution.streaming.MicroBatchExecution.org$apache$spark$sql$execution$streaming$MicroBatchExecution$$runBatch(MicroBatchExecution.scala:534) > at > org.apache.spark.sql.execution.streaming.MicroBatchExecution$$anonfun$runActivatedStream$1$$anonfun$apply$mcZ$sp$1.apply$mcV$sp(MicroBatchExecution.scala:198) > at > org.apache.spark.sql.execution.streaming.MicroBatchExecution$$anonfun$runActivatedStream$1$$anonfun$apply$mcZ$sp$1.apply(MicroBatchExecution.scala:166) > at > org.apache.spark.sql.execution.streaming.MicroBatchExecution$$anonfun$runActivatedStream$1$$anonfun$apply$mcZ$sp$1.apply(MicroBatchExecution.scala:166) > at > org.apache.spark.sql.execution.streaming.ProgressReporter$class.reportTimeTaken(ProgressReporter.scala:351) > at > org.apache.spark.sql.execution.streaming.StreamExecution.reportTimeTaken(StreamExecution.scala:58) > at > org.apache.spark.sql.execution.streaming.MicroBatchExecution$$anonfun$runActivatedStream$1.apply$mcZ$sp(MicroBatchExecution.scala:166) > at > org.apache.spark.sql.execution.streaming.ProcessingTimeExecutor.execute(TriggerExecutor.scala:56) > at > org.apache.spark.sql.execution.streaming.MicroBatchExecution.runActivatedStream(MicroBatchExecution.scala:160) > at >
[jira] [Updated] (HUDI-867) Graphite metrics are throwing IllegalArgumentException on continuous mode
[ https://issues.apache.org/jira/browse/HUDI-867?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Balaji Varadarajan updated HUDI-867: Status: Open (was: New) > Graphite metrics are throwing IllegalArgumentException on continuous mode > - > > Key: HUDI-867 > URL: https://issues.apache.org/jira/browse/HUDI-867 > Project: Apache Hudi (incubating) > Issue Type: Bug > Components: DeltaStreamer >Reporter: João Esteves >Priority: Major > Labels: bug-bash-0.6.0 > > Hello everyone, I am trying to extract Graphite metrics from Hudi using a > Spark Streaming process, but the method that sends metrics is throwing > java.lang.IllegalArgumentException after the first microbatch, like this: > {code:java} > 20/05/06 11:49:25 ERROR Metrics: Failed to send metrics: > java.lang.IllegalArgumentException: A metric named > kafka_hudi.finalize.duration already exists > at > org.apache.hudi.com.codahale.metrics.MetricRegistry.register(MetricRegistry.java:97) > at org.apache.hudi.metrics.Metrics.registerGauge(Metrics.java:83) > at > org.apache.hudi.metrics.HoodieMetrics.updateFinalizeWriteMetrics(HoodieMetrics.java:177) > at > org.apache.hudi.HoodieWriteClient.lambda$finalizeWrite$14(HoodieWriteClient.java:1233) > at org.apache.hudi.common.util.Option.ifPresent(Option.java:96) > at > org.apache.hudi.HoodieWriteClient.finalizeWrite(HoodieWriteClient.java:1231) > at org.apache.hudi.HoodieWriteClient.commit(HoodieWriteClient.java:497) > at org.apache.hudi.HoodieWriteClient.commit(HoodieWriteClient.java:479) > at org.apache.hudi.HoodieWriteClient.commit(HoodieWriteClient.java:470) > at > org.apache.hudi.HoodieSparkSqlWriter$.write(HoodieSparkSqlWriter.scala:152) > at > org.apache.hudi.HoodieStreamingSink$$anonfun$1$$anonfun$2.apply(HoodieStreamingSink.scala:51) > at > org.apache.hudi.HoodieStreamingSink$$anonfun$1$$anonfun$2.apply(HoodieStreamingSink.scala:51) > at scala.util.Try$.apply(Try.scala:192) > at > org.apache.hudi.HoodieStreamingSink$$anonfun$1.apply(HoodieStreamingSink.scala:50) > at > org.apache.hudi.HoodieStreamingSink$$anonfun$1.apply(HoodieStreamingSink.scala:50) > at > org.apache.hudi.HoodieStreamingSink.retry(HoodieStreamingSink.scala:114) > at > org.apache.hudi.HoodieStreamingSink.addBatch(HoodieStreamingSink.scala:49) > at > org.apache.spark.sql.execution.streaming.MicroBatchExecution$$anonfun$org$apache$spark$sql$execution$streaming$MicroBatchExecution$$runBatch$5$$anonfun$apply$17.apply(MicroBatchExecution.scala:537) > at > org.apache.spark.sql.execution.SQLExecution$$anonfun$withNewExecutionId$1.apply(SQLExecution.scala:84) > at > org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:165) > at > org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:74) > at > org.apache.spark.sql.execution.streaming.MicroBatchExecution$$anonfun$org$apache$spark$sql$execution$streaming$MicroBatchExecution$$runBatch$5.apply(MicroBatchExecution.scala:535) > at > org.apache.spark.sql.execution.streaming.ProgressReporter$class.reportTimeTaken(ProgressReporter.scala:351) > at > org.apache.spark.sql.execution.streaming.StreamExecution.reportTimeTaken(StreamExecution.scala:58) > at > org.apache.spark.sql.execution.streaming.MicroBatchExecution.org$apache$spark$sql$execution$streaming$MicroBatchExecution$$runBatch(MicroBatchExecution.scala:534) > at > org.apache.spark.sql.execution.streaming.MicroBatchExecution$$anonfun$runActivatedStream$1$$anonfun$apply$mcZ$sp$1.apply$mcV$sp(MicroBatchExecution.scala:198) > at > org.apache.spark.sql.execution.streaming.MicroBatchExecution$$anonfun$runActivatedStream$1$$anonfun$apply$mcZ$sp$1.apply(MicroBatchExecution.scala:166) > at > org.apache.spark.sql.execution.streaming.MicroBatchExecution$$anonfun$runActivatedStream$1$$anonfun$apply$mcZ$sp$1.apply(MicroBatchExecution.scala:166) > at > org.apache.spark.sql.execution.streaming.ProgressReporter$class.reportTimeTaken(ProgressReporter.scala:351) > at > org.apache.spark.sql.execution.streaming.StreamExecution.reportTimeTaken(StreamExecution.scala:58) > at > org.apache.spark.sql.execution.streaming.MicroBatchExecution$$anonfun$runActivatedStream$1.apply$mcZ$sp(MicroBatchExecution.scala:166) > at > org.apache.spark.sql.execution.streaming.ProcessingTimeExecutor.execute(TriggerExecutor.scala:56) > at > org.apache.spark.sql.execution.streaming.MicroBatchExecution.runActivatedStream(MicroBatchExecution.scala:160) > at >
[jira] [Updated] (HUDI-867) Graphite metrics are throwing IllegalArgumentException on continuous mode
[ https://issues.apache.org/jira/browse/HUDI-867?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] sivabalan narayanan updated HUDI-867: - Labels: bug-bash-0.6.0 (was: ) > Graphite metrics are throwing IllegalArgumentException on continuous mode > - > > Key: HUDI-867 > URL: https://issues.apache.org/jira/browse/HUDI-867 > Project: Apache Hudi (incubating) > Issue Type: Bug > Components: DeltaStreamer >Reporter: João Esteves >Priority: Major > Labels: bug-bash-0.6.0 > > Hello everyone, I am trying to extract Graphite metrics from Hudi using a > Spark Streaming process, but the method that sends metrics is throwing > java.lang.IllegalArgumentException after the first microbatch, like this: > {code:java} > 20/05/06 11:49:25 ERROR Metrics: Failed to send metrics: > java.lang.IllegalArgumentException: A metric named > kafka_hudi.finalize.duration already exists > at > org.apache.hudi.com.codahale.metrics.MetricRegistry.register(MetricRegistry.java:97) > at org.apache.hudi.metrics.Metrics.registerGauge(Metrics.java:83) > at > org.apache.hudi.metrics.HoodieMetrics.updateFinalizeWriteMetrics(HoodieMetrics.java:177) > at > org.apache.hudi.HoodieWriteClient.lambda$finalizeWrite$14(HoodieWriteClient.java:1233) > at org.apache.hudi.common.util.Option.ifPresent(Option.java:96) > at > org.apache.hudi.HoodieWriteClient.finalizeWrite(HoodieWriteClient.java:1231) > at org.apache.hudi.HoodieWriteClient.commit(HoodieWriteClient.java:497) > at org.apache.hudi.HoodieWriteClient.commit(HoodieWriteClient.java:479) > at org.apache.hudi.HoodieWriteClient.commit(HoodieWriteClient.java:470) > at > org.apache.hudi.HoodieSparkSqlWriter$.write(HoodieSparkSqlWriter.scala:152) > at > org.apache.hudi.HoodieStreamingSink$$anonfun$1$$anonfun$2.apply(HoodieStreamingSink.scala:51) > at > org.apache.hudi.HoodieStreamingSink$$anonfun$1$$anonfun$2.apply(HoodieStreamingSink.scala:51) > at scala.util.Try$.apply(Try.scala:192) > at > org.apache.hudi.HoodieStreamingSink$$anonfun$1.apply(HoodieStreamingSink.scala:50) > at > org.apache.hudi.HoodieStreamingSink$$anonfun$1.apply(HoodieStreamingSink.scala:50) > at > org.apache.hudi.HoodieStreamingSink.retry(HoodieStreamingSink.scala:114) > at > org.apache.hudi.HoodieStreamingSink.addBatch(HoodieStreamingSink.scala:49) > at > org.apache.spark.sql.execution.streaming.MicroBatchExecution$$anonfun$org$apache$spark$sql$execution$streaming$MicroBatchExecution$$runBatch$5$$anonfun$apply$17.apply(MicroBatchExecution.scala:537) > at > org.apache.spark.sql.execution.SQLExecution$$anonfun$withNewExecutionId$1.apply(SQLExecution.scala:84) > at > org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:165) > at > org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:74) > at > org.apache.spark.sql.execution.streaming.MicroBatchExecution$$anonfun$org$apache$spark$sql$execution$streaming$MicroBatchExecution$$runBatch$5.apply(MicroBatchExecution.scala:535) > at > org.apache.spark.sql.execution.streaming.ProgressReporter$class.reportTimeTaken(ProgressReporter.scala:351) > at > org.apache.spark.sql.execution.streaming.StreamExecution.reportTimeTaken(StreamExecution.scala:58) > at > org.apache.spark.sql.execution.streaming.MicroBatchExecution.org$apache$spark$sql$execution$streaming$MicroBatchExecution$$runBatch(MicroBatchExecution.scala:534) > at > org.apache.spark.sql.execution.streaming.MicroBatchExecution$$anonfun$runActivatedStream$1$$anonfun$apply$mcZ$sp$1.apply$mcV$sp(MicroBatchExecution.scala:198) > at > org.apache.spark.sql.execution.streaming.MicroBatchExecution$$anonfun$runActivatedStream$1$$anonfun$apply$mcZ$sp$1.apply(MicroBatchExecution.scala:166) > at > org.apache.spark.sql.execution.streaming.MicroBatchExecution$$anonfun$runActivatedStream$1$$anonfun$apply$mcZ$sp$1.apply(MicroBatchExecution.scala:166) > at > org.apache.spark.sql.execution.streaming.ProgressReporter$class.reportTimeTaken(ProgressReporter.scala:351) > at > org.apache.spark.sql.execution.streaming.StreamExecution.reportTimeTaken(StreamExecution.scala:58) > at > org.apache.spark.sql.execution.streaming.MicroBatchExecution$$anonfun$runActivatedStream$1.apply$mcZ$sp(MicroBatchExecution.scala:166) > at > org.apache.spark.sql.execution.streaming.ProcessingTimeExecutor.execute(TriggerExecutor.scala:56) > at > org.apache.spark.sql.execution.streaming.MicroBatchExecution.runActivatedStream(MicroBatchExecution.scala:160) > at >
[jira] [Updated] (HUDI-867) Graphite metrics are throwing IllegalArgumentException on continuous mode
[ https://issues.apache.org/jira/browse/HUDI-867?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] João Esteves updated HUDI-867: -- Summary: Graphite metrics are throwing IllegalArgumentException on continuous mode (was: Graphite metrics are generating IllegalArgumentException on continuous mode) > Graphite metrics are throwing IllegalArgumentException on continuous mode > - > > Key: HUDI-867 > URL: https://issues.apache.org/jira/browse/HUDI-867 > Project: Apache Hudi (incubating) > Issue Type: Bug > Components: DeltaStreamer >Reporter: João Esteves >Priority: Major > > Hello everyone, I am trying to extract Graphite metrics from Hudi using a > Spark Streaming process, but the method that sends metrics is throwing > java.lang.IllegalArgumentException after the first microbatch, like this: > {code:java} > 20/05/06 11:49:25 ERROR Metrics: Failed to send metrics: > java.lang.IllegalArgumentException: A metric named > kafka_hudi.finalize.duration already exists > at > org.apache.hudi.com.codahale.metrics.MetricRegistry.register(MetricRegistry.java:97) > at org.apache.hudi.metrics.Metrics.registerGauge(Metrics.java:83) > at > org.apache.hudi.metrics.HoodieMetrics.updateFinalizeWriteMetrics(HoodieMetrics.java:177) > at > org.apache.hudi.HoodieWriteClient.lambda$finalizeWrite$14(HoodieWriteClient.java:1233) > at org.apache.hudi.common.util.Option.ifPresent(Option.java:96) > at > org.apache.hudi.HoodieWriteClient.finalizeWrite(HoodieWriteClient.java:1231) > at org.apache.hudi.HoodieWriteClient.commit(HoodieWriteClient.java:497) > at org.apache.hudi.HoodieWriteClient.commit(HoodieWriteClient.java:479) > at org.apache.hudi.HoodieWriteClient.commit(HoodieWriteClient.java:470) > at > org.apache.hudi.HoodieSparkSqlWriter$.write(HoodieSparkSqlWriter.scala:152) > at > org.apache.hudi.HoodieStreamingSink$$anonfun$1$$anonfun$2.apply(HoodieStreamingSink.scala:51) > at > org.apache.hudi.HoodieStreamingSink$$anonfun$1$$anonfun$2.apply(HoodieStreamingSink.scala:51) > at scala.util.Try$.apply(Try.scala:192) > at > org.apache.hudi.HoodieStreamingSink$$anonfun$1.apply(HoodieStreamingSink.scala:50) > at > org.apache.hudi.HoodieStreamingSink$$anonfun$1.apply(HoodieStreamingSink.scala:50) > at > org.apache.hudi.HoodieStreamingSink.retry(HoodieStreamingSink.scala:114) > at > org.apache.hudi.HoodieStreamingSink.addBatch(HoodieStreamingSink.scala:49) > at > org.apache.spark.sql.execution.streaming.MicroBatchExecution$$anonfun$org$apache$spark$sql$execution$streaming$MicroBatchExecution$$runBatch$5$$anonfun$apply$17.apply(MicroBatchExecution.scala:537) > at > org.apache.spark.sql.execution.SQLExecution$$anonfun$withNewExecutionId$1.apply(SQLExecution.scala:84) > at > org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:165) > at > org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:74) > at > org.apache.spark.sql.execution.streaming.MicroBatchExecution$$anonfun$org$apache$spark$sql$execution$streaming$MicroBatchExecution$$runBatch$5.apply(MicroBatchExecution.scala:535) > at > org.apache.spark.sql.execution.streaming.ProgressReporter$class.reportTimeTaken(ProgressReporter.scala:351) > at > org.apache.spark.sql.execution.streaming.StreamExecution.reportTimeTaken(StreamExecution.scala:58) > at > org.apache.spark.sql.execution.streaming.MicroBatchExecution.org$apache$spark$sql$execution$streaming$MicroBatchExecution$$runBatch(MicroBatchExecution.scala:534) > at > org.apache.spark.sql.execution.streaming.MicroBatchExecution$$anonfun$runActivatedStream$1$$anonfun$apply$mcZ$sp$1.apply$mcV$sp(MicroBatchExecution.scala:198) > at > org.apache.spark.sql.execution.streaming.MicroBatchExecution$$anonfun$runActivatedStream$1$$anonfun$apply$mcZ$sp$1.apply(MicroBatchExecution.scala:166) > at > org.apache.spark.sql.execution.streaming.MicroBatchExecution$$anonfun$runActivatedStream$1$$anonfun$apply$mcZ$sp$1.apply(MicroBatchExecution.scala:166) > at > org.apache.spark.sql.execution.streaming.ProgressReporter$class.reportTimeTaken(ProgressReporter.scala:351) > at > org.apache.spark.sql.execution.streaming.StreamExecution.reportTimeTaken(StreamExecution.scala:58) > at > org.apache.spark.sql.execution.streaming.MicroBatchExecution$$anonfun$runActivatedStream$1.apply$mcZ$sp(MicroBatchExecution.scala:166) > at > org.apache.spark.sql.execution.streaming.ProcessingTimeExecutor.execute(TriggerExecutor.scala:56) > at > org.apache.spark.sql.execution.streaming.MicroBatchExecution.runActivatedStream(MicroBatchExecution.scala:160) > at >