Hi Dan,

You won't see active frameworks happening until you start running a Spark
job. This is because each Spark job actually launches a new Spark framework
that is scheduling for that single job.

Tim

On Fri, Feb 27, 2015 at 1:39 PM, Dan Dong <[email protected]> wrote:

> Hi, Dick,
>   By Spark daemons I mean Master and Worker process running on master and
> slaves respectively when you run sbin/start-master.sh and
> ./sbin/start-slaves.sh.
>
> Cheers,
> Dan
>
>
> 2015-02-27 15:02 GMT-06:00 Dick Davies <[email protected]>:
>
> What do you mean by spark daemons?
>>
>> the spark shell (or any other spark application) acts as a Mesos
>> framework,
>> so until that's running spark isn't 'on' Mesos.
>>
>> On 27 February 2015 at 16:23, Dan Dong <[email protected]> wrote:
>> > Hi, All,
>> >   When I configured and started spark daemons, why I could not see it in
>> > "Active Frameworks" on Mesos UI(for Hadoop, it will show up immediately
>> when
>> > hadoop daemons started)?. I can see the spark framework on mesos UI only
>> > when I run the spark-shell command interactively. Is it normal?
>> >
>> > export
>> > MESOS_NATIVE_LIBRARY=/home/ubuntu/mesos-0.21.0/build/lib/
>> libmesos-0.21.0.so
>> > export
>> > PROTOBUF_JAR=/home/ubuntu/hadoop-2.5.0-cdh5.2.0/protobuf-java-2.5.0.jar
>> > export MESOS_JAR=/home/ubuntu/hadoop-2.5.0-cdh5.2.0//mesos-0.21.0.jar
>> > export
>> >
>> SPARK_EXECUTOR_URI=hdfs://clus-1:9000/user/ubuntu/spark-1.1.1-bin-hadoop2.4_mesos.tar.gz
>> > export MASTER=mesos://clus-1:5050/mesos
>> >
>> >
>> > Cheers,
>> > Dan
>> >
>>
>
>

Reply via email to