Did you build the spark hadoop 2.2 yarn version or just the regular one? There is no CDH5 build available. We used spark 0.8.1 built with cdh4.2.1-mrv1 before.
We will give it a shot tomorrow and see what happens :) On Mar 25, 2014 9:01 PM, "Bernardo Gomez Palacio" < [email protected]> wrote: > In that setup I am running Spark on Mesos not YARN. > > On Tue, Mar 25, 2014 at 5:55 PM, Gary Malouf <[email protected]> > wrote: > > Thank you for the feedback. Are you running Yarn as a part of this > setup? > > Seems like that would collide with Mesos. We have a job or 2 written in > > MRV1 that from what I understand can still be run with 2.2 somehow... > > On Mar 25, 2014 8:49 PM, "Bernardo Gomez Palacio" < > > [email protected]> wrote: > > > >> Hi Gary. > >> I am currently using > >> Mesos 0.16.0 build for Protobufs 2.5.0 > >> (https://github.com/Guavus/mesos/tree/releases/0.16.0-HDP2x) > >> Hadoop version 2.2 > >> Spark 0.9.0 ( > >> https://github.com/guavus/incubator-spark/tree/guavus/branch-0.9-hdp2.2 > ) > >> > >> Things to know. > >> Hadoop 1.0.4 uses protobufs 2.4.1 > >> Hadoop 2.2 and 2.3 uses protobufs 2.5.0 > >> Mesos < 0.17.0 uses protobufs 2.4.1 > >> Mesos >= 0.17.0 uses protobufs 2.5.0 > >> > >> If you need Mesos 0.15.0 on protobufs 2.5.0 you could try to backport > >> https://github.com/Guavus/mesos/tree/releases/0.16.0-HDP2x but if you > >> keep CDH4 with protobufs 2.4.1 you will run into trouble. > >> > >> Let me know if this help. > >> Bernardo. > >> > >> > >> > >> > >> On Tue, Mar 25, 2014 at 4:57 PM, Gary Malouf <[email protected]> > >> wrote: > >> > For various reasons, our team needs to keep all of our projects on the > >> same > >> > protobuf version. We've now hit a point where we need to upgrade > >> protobuf > >> > from 2.4.1 to 2.5.0 across the board in our projects and dependent > >> > platforms. > >> > > >> > *Current stack*: Mesos 0.15, Chronos, CDH 4.2.1-MRV1, Spark > >> > 0.9-pre-scala-2.10 build off master > >> > > >> > *Ideal stack after upgrade*: Mesos 0.17, Chronos, CDH5 beta2, Spark > 0.9.1 > >> > (hadoop 2.2 build) > >> > > >> > From what we understand, we need a dependency on Hadoop 2.2 to get the > >> > necessary protobuf upgrade. From reading Cloudera's documentation and > >> > multiple google searches, it is not clear to me how we can construct > the > >> > stack to continue to work. > >> > > >> > Has anyone else requested info on getting this combination to work? > From > >> > others we've spoken to, they've basically said that we'll be forced to > >> use > >> > Yarn for Hadoop support in the very near future anyway so we should > >> switch. > >> > Since we colocate Spark with out HDFS nodes, it's hard to see how we > >> would > >> > run both Yarn and Mesos on the same servers. > >> >
