Hi,
I was looking for metrics specifying how many objects ("files") were read /
written when using Spark over S3.
The metrics specified at [
https://spark.apache.org/docs/3.5.1/monitoring.html#component-instance--executor]
do not have objects written / read from s3 metric.
I do see the Hadoop dependency Spark is using to read/write from S3 does
have S3Instrumentation which seems to have rich metrics.
I was wondering is there a place I've missed, to take those read/write
objects count metric ?
Thanks,
Asaf