There is no hadoop-2.5 profile. You can use hadoop-2.4 for 2.4+. This profile already sets protobuf.version to 2.5.0 for this reason. It is already something you can set on the command line as it is read as a Maven build property. It does not pick up an older version because it's somewhere on your system; it uses what is appropriate for your Hadoop version. No changes needed, you just need to follow http://spark.apache.org/docs/latest/hadoop-third-party-distributions.html and set one profile and hadoop version.
On Fri, Dec 12, 2014 at 2:08 PM, kmurph <k.l.mur...@qub.ac.uk> wrote: > I had this problem also with spark 1.1.1. At the time I was using hadoop > 0.20. > > To get around it I installed hadoop 2.5.2, and set the protobuf.version to > 2.5.0 in the build command like so: > mvn -Phadoop-2.5 -Dhadoop.version=2.5.2 -Dprotobuf.version=2.5.0 > -DskipTests clean package > > So I changed spark's pom.xml to read the protobuf.version from the command > line. > If I didn't explicitly set protobuf.version it was picking up an older > version that existed on my filesystem somewhere, > > Karen > > > > -- > View this message in context: > http://apache-spark-user-list.1001560.n3.nabble.com/Spark-1-1-1-Hadoop-2-6-Protobuf-conflict-tp20656p20658.html > Sent from the Apache Spark User List mailing list archive at Nabble.com. > > --------------------------------------------------------------------- > To unsubscribe, e-mail: user-unsubscr...@spark.apache.org > For additional commands, e-mail: user-h...@spark.apache.org > --------------------------------------------------------------------- To unsubscribe, e-mail: user-unsubscr...@spark.apache.org For additional commands, e-mail: user-h...@spark.apache.org