Thank you. That seems to resolve it. On Fri, Nov 6, 2015 at 11:46 PM, Ted Yu <yuzhih...@gmail.com> wrote:
> You mentioned resourcemanager but not nodemanagers. > > I think you need to install Spark on nodes running nodemanagers. > > Cheers > > On Fri, Nov 6, 2015 at 1:32 PM, Kayode Odeyemi <drey...@gmail.com> wrote: > >> Hi, >> >> I have a YARN hadoop setup of 8 nodes (7 datanodes, 1 namenode and >> resourcemaneger). I have Spark setup only on the namenode/resource manager. >> >> Do I need to have Spark installed on the datanodes? >> >> I asked because I'm getting below error when I run a Spark job through >> spark-submit: >> >> Error: Could not find or load main class >> org.apache.spark.deploy.yarn.ExecutorLauncher >> >> I appreciate your help. >> >> Many thanks >> >> >