In that setup I am running Spark on Mesos not YARN.
On Tue, Mar 25, 2014 at 5:55 PM, Gary Malouf <[email protected]> wrote: > Thank you for the feedback. Are you running Yarn as a part of this setup? > Seems like that would collide with Mesos. We have a job or 2 written in > MRV1 that from what I understand can still be run with 2.2 somehow... > On Mar 25, 2014 8:49 PM, "Bernardo Gomez Palacio" < > [email protected]> wrote: > >> Hi Gary. >> I am currently using >> Mesos 0.16.0 build for Protobufs 2.5.0 >> (https://github.com/Guavus/mesos/tree/releases/0.16.0-HDP2x) >> Hadoop version 2.2 >> Spark 0.9.0 ( >> https://github.com/guavus/incubator-spark/tree/guavus/branch-0.9-hdp2.2) >> >> Things to know. >> Hadoop 1.0.4 uses protobufs 2.4.1 >> Hadoop 2.2 and 2.3 uses protobufs 2.5.0 >> Mesos < 0.17.0 uses protobufs 2.4.1 >> Mesos >= 0.17.0 uses protobufs 2.5.0 >> >> If you need Mesos 0.15.0 on protobufs 2.5.0 you could try to backport >> https://github.com/Guavus/mesos/tree/releases/0.16.0-HDP2x but if you >> keep CDH4 with protobufs 2.4.1 you will run into trouble. >> >> Let me know if this help. >> Bernardo. >> >> >> >> >> On Tue, Mar 25, 2014 at 4:57 PM, Gary Malouf <[email protected]> >> wrote: >> > For various reasons, our team needs to keep all of our projects on the >> same >> > protobuf version. We've now hit a point where we need to upgrade >> protobuf >> > from 2.4.1 to 2.5.0 across the board in our projects and dependent >> > platforms. >> > >> > *Current stack*: Mesos 0.15, Chronos, CDH 4.2.1-MRV1, Spark >> > 0.9-pre-scala-2.10 build off master >> > >> > *Ideal stack after upgrade*: Mesos 0.17, Chronos, CDH5 beta2, Spark 0.9.1 >> > (hadoop 2.2 build) >> > >> > From what we understand, we need a dependency on Hadoop 2.2 to get the >> > necessary protobuf upgrade. From reading Cloudera's documentation and >> > multiple google searches, it is not clear to me how we can construct the >> > stack to continue to work. >> > >> > Has anyone else requested info on getting this combination to work? From >> > others we've spoken to, they've basically said that we'll be forced to >> use >> > Yarn for Hadoop support in the very near future anyway so we should >> switch. >> > Since we colocate Spark with out HDFS nodes, it's hard to see how we >> would >> > run both Yarn and Mesos on the same servers. >>
