Update on this... it works when I run the main class with the ./run-example script, but not in any form of scala myjar.jar .
On Mon, Dec 2, 2013 at 7:14 PM, Walrus theCat <[email protected]>wrote: > Anyone have any ideas based on the stack trace? > > Thanks > > > On Sun, Dec 1, 2013 at 9:09 PM, Walrus theCat <[email protected]>wrote: > >> Shouldn't? I imported the new 0.8.0 jars into my build path, and had to >> update my imports accordingly. The only way I upload the spark jars myself >> is that they get packaged into my executable jar. The cluster should have >> the right version based on the flag used to launch it (and it does.) >> >> >> On Fri, Nov 29, 2013 at 10:12 PM, Ashish Rangole <[email protected]>wrote: >> >>> I am sure you have already checked this, any chance the classpath has >>> v 0.7.x jars in it? >>> On Nov 29, 2013 4:40 PM, "Walrus theCat" <[email protected]> wrote: >>> >>>> The "full context" isn't much -- this is the first thing I do in my >>>> main method (assign a value to sc), and it throws this error. >>>> >>>> >>>> On Fri, Nov 29, 2013 at 10:38 AM, Walrus theCat <[email protected] >>>> > wrote: >>>> >>>>> Hi Matei, >>>>> >>>>> Good to hear from you. The stack trace is below. I launched the >>>>> instances with --spark-version=0.8.0 and verified that the version was >>>>> correct by launching spark-shell. Also verified that the version I've got >>>>> in my project is 0.8.0. Nothing else should have changed, as the scripts >>>>> I >>>>> use to set up the classpath and everything is the exact same as I used in >>>>> 0.7.3. >>>>> >>>>> Cheers, >>>>> >>>>> >>>>> java.lang.Exception: Could not find resource path for Web UI: >>>>> org/apache/spark/ui/static >>>>> at >>>>> org.apache.spark.ui.JettyUtils$.createStaticHandler(JettyUtils.scala:89) >>>>> at org.apache.spark.ui.SparkUI.<init>(SparkUI.scala:40) >>>>> at org.apache.spark.SparkContext.<init>(SparkContext.scala:122) >>>>> at walrusthecat.ml.spark.SparkSVM$.main(SparkSVM.scala:16) >>>>> at walrusthecat.ml.spark.SparkSVM.main(SparkSVM.scala) >>>>> at walrusthecat.ml.spark.Main.main(Main.java:7) >>>>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >>>>> at >>>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) >>>>> at >>>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >>>>> at java.lang.reflect.Method.invoke(Method.java:606) >>>>> at >>>>> org.eclipse.jdt.internal.jarinjarloader.JarRsrcLoader.main(JarRsrcLoader.java:58) >>>>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >>>>> at >>>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) >>>>> at >>>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >>>>> at java.lang.reflect.Method.invoke(Method.java:606) >>>>> at >>>>> scala.tools.nsc.util.ScalaClassLoader$$anonfun$run$1.apply(ScalaClassLoader.scala:78) >>>>> at >>>>> scala.tools.nsc.util.ScalaClassLoader$class.asContext(ScalaClassLoader.scala:24) >>>>> at >>>>> scala.tools.nsc.util.ScalaClassLoader$URLClassLoader.asContext(ScalaClassLoader.scala:88) >>>>> at >>>>> scala.tools.nsc.util.ScalaClassLoader$class.run(ScalaClassLoader.scala:78) >>>>> at >>>>> scala.tools.nsc.util.ScalaClassLoader$URLClassLoader.run(ScalaClassLoader.scala:101) >>>>> at scala.tools.nsc.ObjectRunner$.run(ObjectRunner.scala:33) >>>>> at scala.tools.nsc.ObjectRunner$.runAndCatch(ObjectRunner.scala:40) >>>>> at >>>>> scala.tools.nsc.MainGenericRunner.runTarget$1(MainGenericRunner.scala:60) >>>>> at >>>>> scala.tools.nsc.MainGenericRunner.process(MainGenericRunner.scala:80) >>>>> at >>>>> scala.tools.nsc.MainGenericRunner$.main(MainGenericRunner.scala:89) >>>>> at scala.tools.nsc.MainGenericRunner.main(MainGenericRunner.scala) >>>>> >>>>> >>>>> >>>>> >>>>> On Wed, Nov 27, 2013 at 6:15 PM, Matei Zaharia < >>>>> [email protected]> wrote: >>>>> >>>>>> Sorry, what’s the full context for this? Do you have a stack trace? >>>>>> My guess is that Spark isn’t on your classpath, or maybe you only have an >>>>>> old version of it on there. >>>>>> >>>>>> Matei >>>>>> >>>>>> On Nov 27, 2013, at 6:04 PM, Walrus theCat <[email protected]> >>>>>> wrote: >>>>>> >>>>>> To clarify, I just undid that "var... field.." thing described above, >>>>>> and it throws the same error. >>>>>> >>>>>> >>>>>> On Wed, Nov 27, 2013 at 5:53 PM, Walrus theCat < >>>>>> [email protected]> wrote: >>>>>> >>>>>>> Hi all, >>>>>>> >>>>>>> This exception gets thrown when I assign a value to the variable >>>>>>> holding my SparkContext. I initialize it as a var holding a null value >>>>>>> (so >>>>>>> it can be a field), and then give it a value in my main method. This >>>>>>> worked with the previous version of Spark, but is not working on Spark >>>>>>> 0.8.0. >>>>>>> >>>>>>> Dankeschöen, >>>>>>> >>>>>>> Walrus theCat >>>>>>> >>>>>> >>>>>> >>>>>> >>>>> >>>> >> >
