Hi,
when I analyze and debug our Spark batch jobs executions it's a pain to
find out how blocks in Spark UI Jobs/SQL tab correspond to the actual
Scala code that we write and how much time they take. Would there be a
way to somehow instruct compiler or something and get this information
into Spark UI?
At the moment linking Spark UI elements with our code is a guess work
driven by adding and removing lines of code and reruning the job, which
is tedious. A possibility to make our life easier e.g. by running Spark
jobs in dedicated debug mode where this information would be available
would be greatly appreciated. (Though I don't know whether it's possible
at all).
Thanks,
Michal
---------------------------------------------------------------------
To unsubscribe e-mail: dev-unsubscr...@spark.apache.org