Hi! Again, thanks a lot for advice: I'll have a look!
Best, Aurelien Le mar. 7 sept. 2021 à 20:36, Haryani, Akshay <akshay.hary...@hpe.com> a écrit : > For custom metrics, you can take a look at Groupon’s spar metrics: > https://github.com/groupon/spark-metrics > > > > It is supported on spark 2.x. Alternatively, you can create a custom > source (extending source trait), enable the sink and register the custom > source to get the metrics. Some useful links for this approach: > > https://gist.github.com/ibuenros/9b94736c2bad2f4b8e23 > > https://kb.databricks.com/metrics/spark-metrics.html > > > http://mail-archives.us.apache.org/mod_mbox/spark-user/201501.mbox/%3CCAE50=dq+6tdx9VNVM3ctBMWPLDPbUAacO3aN3L8x38zg=xb...@mail.gmail.com%3E > > > > I hope these help. > > -- > > Thanks & Regards, > > Akshay Haryani > > > > *From: *Aurélien Mazoyer <aurel...@aepsilon.com> > *Date: *Monday, September 6, 2021 at 5:47 AM > *To: *Haryani, Akshay <akshay.hary...@hpe.com> > *Cc: *user@spark.apache.org <user@spark.apache.org> > *Subject: *Re: Get application metric from Spark job > > Hi Akshay, > > > > Thank you for your reply. Sounds like a good idea, but I unfortunately > have a 2.6 cluster. Do you know if there would be another solution that > would run on 2.6 or if I have no other choice than migrating to 3? > > > > Regards, > > > > Aurélien > > > > Le jeu. 2 sept. 2021 à 20:12, Haryani, Akshay <akshay.hary...@hpe.com> a > écrit : > > Hi Aurélien, > > > > Spark has endpoints to expose the spark application metrics. These > endpoints can be used as a rest API. You can read more about these here: > https://spark.apache.org/docs/3.1.1/monitoring.html#rest-api > > > > Additionally, > > If you want to build your own custom metrics, you can explore spark custom > plugins. Using a custom plugin, you can track your own custom metrics and > plug it into the spark metrics system. Please note plugins are supported > on spark versions above 3.0. > > > > > > -- > > Thanks & Regards, > > Akshay Haryani > > > > *From: *Aurélien Mazoyer <aurel...@aepsilon.com> > *Date: *Thursday, September 2, 2021 at 8:36 AM > *To: *user@spark.apache.org <user@spark.apache.org> > *Subject: *Get application metric from Spark job > > Hi community, > > > > I would like to collect information about the execution of a Spark job > while it is running. Could I define some kind of application metrics (such > as a counter that would be incremented in my code) that I could retrieve > regularly while the job is running? > > > Thank you for help, > > > > Aurelien > >