Hi Jey,

The work directory seems to be empty, and the same for the logs
directory. I can only see the job's Web UI and it says that it failed.
In the Errors column it shows the same error as appears in the Python
console:

org.apache.spark.SparkException (org.apache.spark.SparkException: Python
worker exited unexpectedly (crashed))

Does this mean the Python workers crash right away when they're started?

Thanks,
Michal


On 21/11/13 23:48, [email protected] wrote:
> Hi Michal,
> 
> Try taking a look at the stderr output of the failed executor. You can
> do this by logging into the Spark worker that had the failed executor,
> and look for the executor ID in the "work" subdirectory. You should
> see a file named "stderr" that has the error messages from the failed
> executor.
> 
> Hope that helps,
> -Jey
> 
> On Thu, Nov 21, 2013 at 1:01 PM, Michal Romaniuk
> <[email protected]> wrote:
>> > Hi,
>> >
>> > I'm trying to run a Python job that keeps failing. It's a map followed
>> > by collect, and works fine if I use Python's built-in map instead of Spark.
>> >
>> > I tried to replace the mapping function with an identity (lambda x: x)
>> > and that works fine with Spark, so Spark seems to be configured correctly.
>> >
>> > The error I get is:
>> >
>> > org.apache.spark.SparkException (org.apache.spark.SparkException: Python
>> > worker exited unexpectedly (crashed))
>> >
>> > The problem is that I can't see what went wrong in the Python code. Any
>> > ideas on how to debug this?
>> >
>> > Thanks,
>> > Michal
>> >

Reply via email to