ueshin opened a new pull request #33273:
URL: https://github.com/apache/spark/pull/33273


   ### What changes were proposed in this pull request?
   
   Try to capture the error message from the `faulthandler` when the Python 
worker crashes.
   
   ### Why are the changes needed?
   
   Currently, we just see an error message saying `"exited unexpectedly 
(crashed)"` when the UDFs causes the Python worker to crash by like 
segmentation fault.
   We should take advantage of 
[`faulthandler`](https://docs.python.org/3/library/faulthandler.html) and try 
to capture the error message from the `faulthandler`.
   
   ### Does this PR introduce _any_ user-facing change?
   
   Yes, when a Spark config `spark.python.worker.faulthandler.enabled` is 
`true`, the stack trace will be seen in the error message when the Python 
worker crashes.
   
   ```py
   >>> def f():
   ...   import ctypes
   ...   ctypes.string_at(0)
   ...
   >>> sc.parallelize([1]).map(lambda x: f()).count()
   ```
   
   ```
   org.apache.spark.SparkException: Python worker exited unexpectedly 
(crashed): Fatal Python error: Segmentation fault
   
   Current thread 0x000000010965b5c0 (most recent call first):
     File 
"/Users/ueshin/miniconda/envs/pyspark_3.9/lib/python3.9/ctypes/__init__.py", 
line 525 in string_at
     File "<stdin>", line 3 in f
     File "<stdin>", line 1 in <lambda>
   ...
   ```
   
   ### How was this patch tested?
   
   Added some tests, and manually.
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]



---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to