you can check from here 
http://spark.incubator.apache.org/docs/latest/cluster-overview.html

the driver is more like a per-application task scheduler whereas the master is 
a resource manager to allocate resources to applications and maintain workers

with the command you use, the driver will run on the worker nodes 

Best, 

-- 
Nan Zhu


On Tuesday, February 18, 2014 at 10:32 PM, David Thomas wrote:

> I was using that exact command line (./bin/spark-class 
> org.apache.spark.deploy.Client launch) to launch the app within the cluster. 
> Now I was running this from the master node, but the Worker column shows just 
> one worker in the cluster. So what is the significance of that column? 
> Shouldn't the driver be running on the master node?
> 
> 
> On Tue, Feb 18, 2014 at 8:31 PM, Nan Zhu <zhunanmcg...@gmail.com 
> (mailto:zhunanmcg...@gmail.com)> wrote:
> > the driver is running on the machine where you run command like 
> > ./spark-shell 
> > 
> > but in 0.9, you can run in-cluster driver 
> > http://spark.incubator.apache.org/docs/latest/spark-standalone.html#launching-applications-inside-the-cluster
> >  
> > 
> > Best, 
> > 
> > -- 
> > Nan Zhu
> > 
> > 
> > On Tuesday, February 18, 2014 at 10:06 PM, David Thomas wrote:
> > 
> > > On the spark web UI, there are two tables - Running Applications and 
> > > Running Drivers. What is the difference between these two? 
> > > 
> > > When I submit an application to run on the cluster, I can see the driver 
> > > table getting populated, but why is only one worker listed under Worker 
> > > column? What is the role of this single worker? Shouldn't the driver 
> > > program be running on the master node?
> > 
> 

Reply via email to