Spark leverages the hadoop S3A connector,
https://hadoop.apache.org/docs/current/hadoop-aws/tools/hadoop-aws/connecting.html
Specifics of S3 metrics are documented with AWS,
https://docs.aws.amazon.com/AmazonS3/latest/userguide/metrics-dimensions.html
Hope this helps.

Best Regards
Soumasish Goswami
in: www.linkedin.com/in/soumasish
# (415) 530-0405

   -



On Mon, Dec 23, 2024 at 1:11 AM Asaf Mesika <asaf.mes...@gmail.com> wrote:

> Hi,
>
> I was looking for metrics specifying how many objects ("files") were read
> / written when using Spark over S3.
>
> The metrics specified at [
> https://spark.apache.org/docs/3.5.1/monitoring.html#component-instance--executor]
> do not have objects written / read from s3 metric.
>
> I do see the Hadoop dependency Spark is using to read/write from S3 does
> have S3Instrumentation which seems to have rich metrics.
>
> I was wondering is there a place I've missed, to take those read/write
> objects count metric ?
>
>
> Thanks,
>
> Asaf
>

Reply via email to