Re: Run spark with hadoop snapshot

2016-11-19 Thread Luke Miner
Thanks! Should I do it from the spark build environment?

On Sat, Nov 19, 2016 at 4:48 AM, Steve Loughran 
wrote:

> I'd recommend you build a fill spark release with the new hadoop version;
> you should have built that locally earlier the same day (so that ivy/maven
> pick up the snapshot)
>
>
> dev/make-distribution.sh -Pyarn,hadoop-2.7,hive -Dhadoop.version=2.9.0-
> SNAPSHOT;
>
>
>
> > On 18 Nov 2016, at 19:31, lminer  wrote:
> >
> > I'm trying to figure out how to run spark with a snapshot of Hadoop 2.8
> that
> > I built myself. I'm unclear on the configuration needed to get spark to
> work
> > with the snapshot.
> >
> > I'm running spark on mesos. Per the spark documentation, I run
> spark-submit
> > as follows using the `spark-2.0.2-bin-without-hadoop`, but spark doesn't
> > appear to be finding hadoop 2.8.
> >
> >export SPARK_DIST_CLASSPATH=$(/path/to/hadoop2.8/bin/hadoop
> classpath)
> >spark-submit --verbose --master mesos://$MASTER_HOST/mesos
> >
> > I get the error:
> >
> >Exception in thread "main" java.lang.NoClassDefFoundError:
> > org/apache/hadoop/fs/FSDataInputStream
> >at
> > org.apache.spark.deploy.SparkSubmitArguments.handle(
> SparkSubmitArguments.scala:403)
> >at
> > org.apache.spark.launcher.SparkSubmitOptionParser.parse(
> SparkSubmitOptionParser.java:163)
> >at
> > org.apache.spark.deploy.SparkSubmitArguments.(
> SparkSubmitArguments.scala:98)
> >at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:117)
> >at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
> > Caused by: java.lang.ClassNotFoundException:
> > org.apache.hadoop.fs.FSDataInputStream
> >at java.net.URLClassLoader$1.run(URLClassLoader.java:366)
> >at java.net.URLClassLoader$1.run(URLClassLoader.java:355)
> >at java.security.AccessController.doPrivileged(Native Method)
> >at java.net.URLClassLoader.findClass(URLClassLoader.java:354)
> >at java.lang.ClassLoader.loadClass(ClassLoader.java:425)
> >at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308)
> >at java.lang.ClassLoader.loadClass(ClassLoader.java:358)
> >... 5 more
> >
> > Any ideas on the proper configuration?
> >
> >
> >
> > --
> > View this message in context: http://apache-spark-user-list.
> 1001560.n3.nabble.com/Run-spark-with-hadoop-snapshot-tp28105.html
> > Sent from the Apache Spark User List mailing list archive at Nabble.com.
> >
> > -
> > To unsubscribe e-mail: user-unsubscr...@spark.apache.org
> >
> >
>
>


Re: Run spark with hadoop snapshot

2016-11-19 Thread Steve Loughran
I'd recommend you build a fill spark release with the new hadoop version; you 
should have built that locally earlier the same day (so that ivy/maven pick up 
the snapshot)


dev/make-distribution.sh -Pyarn,hadoop-2.7,hive -Dhadoop.version=2.9.0-SNAPSHOT;



> On 18 Nov 2016, at 19:31, lminer  wrote:
> 
> I'm trying to figure out how to run spark with a snapshot of Hadoop 2.8 that
> I built myself. I'm unclear on the configuration needed to get spark to work
> with the snapshot.
> 
> I'm running spark on mesos. Per the spark documentation, I run spark-submit
> as follows using the `spark-2.0.2-bin-without-hadoop`, but spark doesn't
> appear to be finding hadoop 2.8.
> 
>export SPARK_DIST_CLASSPATH=$(/path/to/hadoop2.8/bin/hadoop classpath)
>spark-submit --verbose --master mesos://$MASTER_HOST/mesos
> 
> I get the error:
> 
>Exception in thread "main" java.lang.NoClassDefFoundError:
> org/apache/hadoop/fs/FSDataInputStream
>at
> org.apache.spark.deploy.SparkSubmitArguments.handle(SparkSubmitArguments.scala:403)
>at
> org.apache.spark.launcher.SparkSubmitOptionParser.parse(SparkSubmitOptionParser.java:163)
>at
> org.apache.spark.deploy.SparkSubmitArguments.(SparkSubmitArguments.scala:98)
>at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:117)
>at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
> Caused by: java.lang.ClassNotFoundException:
> org.apache.hadoop.fs.FSDataInputStream
>at java.net.URLClassLoader$1.run(URLClassLoader.java:366)
>at java.net.URLClassLoader$1.run(URLClassLoader.java:355)
>at java.security.AccessController.doPrivileged(Native Method)
>at java.net.URLClassLoader.findClass(URLClassLoader.java:354)
>at java.lang.ClassLoader.loadClass(ClassLoader.java:425)
>at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308)
>at java.lang.ClassLoader.loadClass(ClassLoader.java:358)
>... 5 more
> 
> Any ideas on the proper configuration?
> 
> 
> 
> --
> View this message in context: 
> http://apache-spark-user-list.1001560.n3.nabble.com/Run-spark-with-hadoop-snapshot-tp28105.html
> Sent from the Apache Spark User List mailing list archive at Nabble.com.
> 
> -
> To unsubscribe e-mail: user-unsubscr...@spark.apache.org
> 
> 


-
To unsubscribe e-mail: user-unsubscr...@spark.apache.org



Run spark with hadoop snapshot

2016-11-18 Thread lminer
I'm trying to figure out how to run spark with a snapshot of Hadoop 2.8 that
I built myself. I'm unclear on the configuration needed to get spark to work
with the snapshot.

I'm running spark on mesos. Per the spark documentation, I run spark-submit
as follows using the `spark-2.0.2-bin-without-hadoop`, but spark doesn't
appear to be finding hadoop 2.8.

export SPARK_DIST_CLASSPATH=$(/path/to/hadoop2.8/bin/hadoop classpath)
spark-submit --verbose --master mesos://$MASTER_HOST/mesos

I get the error:

Exception in thread "main" java.lang.NoClassDefFoundError:
org/apache/hadoop/fs/FSDataInputStream
at
org.apache.spark.deploy.SparkSubmitArguments.handle(SparkSubmitArguments.scala:403)
at
org.apache.spark.launcher.SparkSubmitOptionParser.parse(SparkSubmitOptionParser.java:163)
at
org.apache.spark.deploy.SparkSubmitArguments.(SparkSubmitArguments.scala:98)
at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:117)
at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
Caused by: java.lang.ClassNotFoundException:
org.apache.hadoop.fs.FSDataInputStream
at java.net.URLClassLoader$1.run(URLClassLoader.java:366)
at java.net.URLClassLoader$1.run(URLClassLoader.java:355)
at java.security.AccessController.doPrivileged(Native Method)
at java.net.URLClassLoader.findClass(URLClassLoader.java:354)
at java.lang.ClassLoader.loadClass(ClassLoader.java:425)
at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308)
at java.lang.ClassLoader.loadClass(ClassLoader.java:358)
... 5 more

Any ideas on the proper configuration?



--
View this message in context: 
http://apache-spark-user-list.1001560.n3.nabble.com/Run-spark-with-hadoop-snapshot-tp28105.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.

-
To unsubscribe e-mail: user-unsubscr...@spark.apache.org