I built Spark for hadoop 2.2 with Mesos, make sure that the YARN env var is
disabled. I updated the Mesos dep to point to a Mesos version that uses
protobufs 2.5.0. That should work with CDH5.



On Tuesday, March 25, 2014, Gary Malouf <[email protected]> wrote:

> Did you build the spark hadoop 2.2 yarn version or just the regular one?
> There is  no CDH5 build available.  We used spark 0.8.1 built with
> cdh4.2.1-mrv1 before.
>
> We will give it a shot tomorrow and see what happens :)
> On Mar 25, 2014 9:01 PM, "Bernardo Gomez Palacio" <
> [email protected] <javascript:;>> wrote:
>
> > In that setup I am running Spark on Mesos not YARN.
> >
> > On Tue, Mar 25, 2014 at 5:55 PM, Gary Malouf 
> > <[email protected]<javascript:;>
> >
> > wrote:
> > > Thank you for the feedback.  Are you running Yarn as a part of this
> > setup?
> > > Seems like that would collide with Mesos.  We have a job or 2 written
> in
> > > MRV1 that from what I understand can still be run with 2.2 somehow...
> > > On Mar 25, 2014 8:49 PM, "Bernardo Gomez Palacio" <
> > > [email protected] <javascript:;>> wrote:
> > >
> > >> Hi Gary.
> > >> I am currently using
> > >> Mesos 0.16.0 build for Protobufs 2.5.0
> > >> (https://github.com/Guavus/mesos/tree/releases/0.16.0-HDP2x)
> > >> Hadoop version 2.2
> > >> Spark 0.9.0 (
> > >>
> https://github.com/guavus/incubator-spark/tree/guavus/branch-0.9-hdp2.2
> > )
> > >>
> > >> Things to know.
> > >> Hadoop 1.0.4 uses protobufs 2.4.1
> > >> Hadoop 2.2 and 2.3 uses protobufs 2.5.0
> > >> Mesos < 0.17.0 uses protobufs 2.4.1
> > >> Mesos >= 0.17.0 uses protobufs 2.5.0
> > >>
> > >> If you need Mesos 0.15.0 on protobufs 2.5.0 you could try to backport
> > >> https://github.com/Guavus/mesos/tree/releases/0.16.0-HDP2x but if you
> > >> keep CDH4 with protobufs 2.4.1 you will run into trouble.
> > >>
> > >> Let me know if this help.
> > >> Bernardo.
> > >>
> > >>
> > >>
> > >>
> > >> On Tue, Mar 25, 2014 at 4:57 PM, Gary Malouf 
> > >> <[email protected]<javascript:;>
> >
> > >> wrote:
> > >> > For various reasons, our team needs to keep all of our projects on
> the
> > >> same
> > >> > protobuf version.  We've now hit a point where we need to upgrade
> > >> protobuf
> > >> > from 2.4.1 to 2.5.0 across the board in our projects and dependent
> > >> > platforms.
> > >> >
> > >> > *Current stack*: Mesos 0.15, Chronos, CDH 4.2.1-MRV1, Spark
> > >> > 0.9-pre-scala-2.10 build off master
> > >> >
> > >> > *Ideal stack after upgrade*: Mesos 0.17, Chronos, CDH5 beta2, Spark
> > 0.9.1
> > >> > (hadoop 2.2 build)
> > >> >
> > >> > From what we understand, we need a dependency on Hadoop 2.2 to get
> the
> > >> > necessary protobuf upgrade.  From reading Cloudera's documentation
> and
> > >> > multiple google searches, it is not clear to me how we can construct
> > the
> > >> > stack to continue to work.
> > >> >
> > >> > Has anyone else requested info on getting this combination to work?
> >  From
> > >> > others we've spoken to, they've basically said that we'll be forced
> to
> > >> use
> > >> > Yarn for Hadoop support in the very near future anyway so we should
> > >> switch.
> > >> >  Since we colocate Spark with out HDFS nodes, it's hard to see how
> we
> > >> would
> > >> > run both Yarn and Mesos on the same servers.
> > >>
> >
>

Reply via email to