Hey Egor,

Have you checked the AM logs? My guess is that it threw an exception or
something such that no executors (not even the initial set) have registered
with your driver. You may already know this, but you can go to the http://<RM
address>:8088 page and click into the application to access this.
Alternatively you could run "yarn logs -applicationId <appId>" after
quitting the application.

-Andrew

2014-11-14 14:23 GMT-08:00 Sandy Ryza <sandy.r...@cloudera.com>:

> That would be helpful as well.  Can you confirm that when you try it with
> dynamic allocation the cluster has free resources?
>
> On Fri, Nov 14, 2014 at 12:17 PM, Egor Pahomov <pahomov.e...@gmail.com>
> wrote:
>
>> It's successful without dynamic allocation. I can provide spark log for
>> that scenario if it can help.
>>
>> 2014-11-14 21:36 GMT+02:00 Sandy Ryza <sandy.r...@cloudera.com>:
>>
>>> Hi Egor,
>>>
>>> Is it successful without dynamic allocation? From your log, it looks
>>> like the job is unable to acquire resources from YARN, which could be
>>> because other jobs are using up all the resources.
>>>
>>> -Sandy
>>>
>>> On Fri, Nov 14, 2014 at 11:32 AM, Egor Pahomov <pahomov.e...@gmail.com>
>>> wrote:
>>>
>>>> Hi.
>>>> I execute ipython notebook + pyspark with
>>>> spark.dynamicAllocation.enabled = true. Task never ends.
>>>> Code:
>>>>
>>>> import sys
>>>> from random import random
>>>> from operator import add
>>>> partitions = 10
>>>> n = 100000 * partitions
>>>>
>>>> def f(_):
>>>>     x = random() * 2 - 1
>>>>     y = random() * 2 - 1
>>>>     return 1 if x ** 2 + y ** 2 < 1 else 0
>>>>
>>>> count = sc.parallelize(xrange(1, n + 1), partitions).map(f).reduce(add)
>>>> print "Pi is roughly %f" % (4.0 * count / n)
>>>>
>>>>
>>>>
>>>> Run notebook:
>>>>
>>>> IPYTHON_ARGS="notebook --profile=ydf --port $IPYTHON_PORT --port-retries=0 
>>>> --ip='*' --no-browser"
>>>> pyspark \
>>>>         --verbose \
>>>>         --master yarn-client \
>>>>         --conf spark.driver.port=$((RANDOM_PORT + 2)) \
>>>>         --conf spark.broadcast.port=$((RANDOM_PORT + 3)) \
>>>>         --conf spark.replClassServer.port=$((RANDOM_PORT + 4)) \
>>>>         --conf spark.blockManager.port=$((RANDOM_PORT + 5)) \
>>>>         --conf spark.executor.port=$((RANDOM_PORT + 6)) \
>>>>         --conf spark.fileserver.port=$((RANDOM_PORT + 7)) \
>>>>         --conf spark.shuffle.service.enabled=true \
>>>>         --conf spark.dynamicAllocation.enabled=true \
>>>>         --conf spark.dynamicAllocation.minExecutors=1 \
>>>>         --conf spark.dynamicAllocation.maxExecutors=10 \
>>>>         --conf spark.ui.port=$SPARK_UI_PORT
>>>>
>>>>
>>>> Spark/Ipython log is in attachment.
>>>>
>>>> --
>>>>
>>>>
>>>>
>>>> *Sincerely yoursEgor PakhomovScala Developer, Yandex*
>>>>
>>>>
>>>> ---------------------------------------------------------------------
>>>> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
>>>> For additional commands, e-mail: user-h...@spark.apache.org
>>>>
>>>
>>>
>>
>>
>> --
>>
>>
>>
>> *Sincerely yoursEgor PakhomovScala Developer, Yandex*
>>
>
>

Reply via email to