Thanks for the quick response!

spark-shell is indeed using yarn-client. I forgot to mention that I also
have "spark.master yarn-client" in my spark-defaults.conf file too.

The working spark-shell and my non-working example application both display
spark.scheduler.mode=FIFO on the Spark UI. Is that what you are asking
about? I haven't actually messed around with different scheduler modes yet.

One more thing I should mention is that the YARN ResourceManager tells me
the following on my 5-node cluster, with one node being the master and not
running a NodeManager:
Memory Used: 1.50 GB (this is the running ApplicationMaster that's waiting
and waiting for the executors to start up)
Memory Total: 45 GB (11.25 from each of the 4 slave nodes)
VCores Used: 1
VCores Total: 32
Active Nodes: 4

~ Jonathan

On Wed, Sep 23, 2015 at 6:10 PM, Andrew Duffy <andrewedu...@gmail.com>
wrote:

> What pool is the spark shell being put into? (You can see this through the
> YARN UI under scheduler)
>
> Are you certain you're starting spark-shell up on YARN? By default it uses
> a local spark executor, so if it "just works" then it's because it's not
> using dynamic allocation.
>
>
> On Wed, Sep 23, 2015 at 18:04 Jonathan Kelly <jonathaka...@gmail.com>
> wrote:
>
>> I'm running into a problem with YARN dynamicAllocation on Spark 1.5.0
>> after using it successfully on an identically configured cluster with Spark
>> 1.4.1.
>>
>> I'm getting the dreaded warning "YarnClusterScheduler: Initial job has
>> not accepted any resources; check your cluster UI to ensure that workers
>> are registered and have sufficient resources", though there's nothing else
>> running on my cluster, and the nodes should have plenty of resources to run
>> my application.
>>
>> Here are the applicable properties in spark-defaults.conf:
>> spark.dynamicAllocation.enabled  true
>> spark.dynamicAllocation.minExecutors 1
>> spark.shuffle.service.enabled true
>>
>> When trying out my example application (just the JavaWordCount example
>> that comes with Spark), I had not actually set spark.executor.memory or any
>> CPU core-related properties, but setting the spark.executor.memory to a low
>> value like 64m doesn't help either.
>>
>> I've tried a 5-node cluster and 1-node cluster of m3.xlarges, so each
>> node has 15.0GB and 4 cores.
>>
>> I've also tried both yarn-cluster and yarn-client mode and get the same
>> behavior for both, except that for yarn-client mode the application never
>> even shows up in the YARN ResourceManager. However, spark-shell seems to
>> work just fine (when I run commands, it starts up executors dynamically
>> just fine), which makes no sense to me.
>>
>> What settings/logs should I look at to debug this, and what more
>> information can I provide? Your help would be very much appreciated!
>>
>> Thanks,
>> Jonathan
>>
>

Reply via email to