Hi,
when I analyze and debug our Spark batch jobs executions it's a pain to find out how blocks in Spark UI Jobs/SQL tab correspond to the actual Scala code that we write and how much time they take. Would there be a way to somehow instruct compiler or something and get this information into Spark UI?

At the moment linking Spark UI elements with our code is a guess work driven by adding and removing lines of code and reruning the job, which is tedious. A possibility to make our life easier e.g. by running Spark jobs in dedicated debug mode where this information would be available would be greatly appreciated. (Though I don't know whether it's possible at all).

Thanks,
Michal

---------------------------------------------------------------------
To unsubscribe e-mail: dev-unsubscr...@spark.apache.org

Reply via email to