Sorry I have a typo. 

Which means spark does not use yarn or mesos in standalone mode...



> On Jun 11, 2016, at 14:35, Mich Talebzadeh <mich.talebza...@gmail.com> wrote:
> 
> Hi Gavin,
> 
> I believe in standalone mode a simple cluster manager is included with Spark 
> that makes it easy to set up a cluster. It does not rely on YARN or Mesos.
> 
> In summary this is from my notes:
> 
> Spark Local - Spark runs on the local host. This is the simplest set up and 
> best suited for learners who want to understand different concepts of Spark 
> and those performing unit testing.
> Spark Standalone – a simple cluster manager included with Spark that makes it 
> easy to set up a cluster.
> YARN Cluster Mode, the Spark driver runs inside an application master process 
> which is managed by YARN on the cluster, and the client can go away after 
> initiating the application.
> Mesos. I have not used it so cannot comment
> YARN Client Mode, the driver runs in the client process, and the application 
> master is only used for requesting resources from YARN. Unlike Local or Spark 
> standalone modes, in which the master’s address is specified in the --master 
> parameter, in YARN mode the ResourceManager’s address is picked up from the 
> Hadoop configuration. Thus, the --master parameter is yarn
> 
> HTH
> 
> 
> 
> 
> Dr Mich Talebzadeh
>  
> LinkedIn  
> https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw
>  
> http://talebzadehmich.wordpress.com
>  
> 
>> On 11 June 2016 at 22:26, Gavin Yue <yue.yuany...@gmail.com> wrote:
>> The standalone mode is against Yarn mode or Mesos mode, which means spark 
>> uses Yarn or Mesos as cluster managements. 
>> 
>> Local mode is actually a standalone mode which everything runs on the single 
>> local machine instead of remote clusters.
>> 
>> That is my understanding. 
>> 
>> 
>>> On Sat, Jun 11, 2016 at 12:40 PM, Ashok Kumar 
>>> <ashok34...@yahoo.com.invalid> wrote:
>>> Thank you for grateful
>>> 
>>> I know I can start spark-shell by launching the shell itself
>>> 
>>> spark-shell 
>>> 
>>> Now I know that in standalone mode I can also connect to master
>>> 
>>> spark-shell --master spark://<HOST>:7077
>>> 
>>> My point is what are the differences between these two start-up modes for 
>>> spark-shell? If I start spark-shell and connect to master what performance 
>>> gain will I get if any or it does not matter. Is it the same as for 
>>> spark-submit 
>>> 
>>> regards
>>> 
>>> 
>>> On Saturday, 11 June 2016, 19:39, Mohammad Tariq <donta...@gmail.com> wrote:
>>> 
>>> 
>>> Hi Ashok,
>>> 
>>> In local mode all the processes run inside a single jvm, whereas in 
>>> standalone mode we have separate master and worker processes running in 
>>> their own jvms.
>>> 
>>> To quickly test your code from within your IDE you could probable use the 
>>> local mode. However, to get a real feel of how Spark operates I would 
>>> suggest you to have a standalone setup as well. It's just the matter of 
>>> launching a standalone cluster either manually(by starting a master and 
>>> workers by hand), or by using the launch scripts provided with Spark 
>>> package. 
>>> 
>>> You can find more on this here.
>>> 
>>> HTH
>>> 
>>>  
>>> 
>>> Tariq, Mohammad
>>> about.me/mti
>>> 
>>> 
>>>                             
>>>  
>>> 
>>> On Sat, Jun 11, 2016 at 11:38 PM, Ashok Kumar 
>>> <ashok34...@yahoo.com.invalid> wrote:
>>> Hi,
>>> 
>>> What is the difference between running Spark in Local mode or standalone 
>>> mode?
>>> 
>>> Are they the same. If they are not which is best suited for non prod work.
>>> 
>>> I am also aware that one can run Spark in Yarn mode as well.
>>> 
>>> Thanks
> 

Reply via email to