To determine if this is a Windows vs. other configuration, can you just try
to call the Spark-class.cmd SparkSubmit without actually referencing the
Hadoop or Thrift server classes?


On Tue Nov 25 2014 at 5:42:09 PM Judy Nash <judyn...@exchange.microsoft.com>
wrote:

>  I traced the code and used the following to call:
>
> Spark-class.cmd org.apache.spark.deploy.SparkSubmit --class
> org.apache.spark.sql.hive.thriftserver.HiveThriftServer2 spark-internal
> --hiveconf hive.server2.thrift.port=10000
>
>
>
> The issue ended up to be much more fundamental however. Spark doesn’t work
> at all in configuration below. When open spark-shell, it fails with the
> same ClassNotFound error.
>
> Now I wonder if this is a windows-only issue or the hive/Hadoop
> configuration that is having this problem.
>
>
>
> *From:* Cheng Lian [mailto:lian.cs....@gmail.com]
> *Sent:* Tuesday, November 25, 2014 1:50 AM
>
>
> *To:* Judy Nash; u...@spark.incubator.apache.org
> *Subject:* Re: latest Spark 1.2 thrift server fail with
> NoClassDefFoundError on Guava
>
>
>
> Oh so you're using Windows. What command are you using to start the Thrift
> server then?
>
> On 11/25/14 4:25 PM, Judy Nash wrote:
>
> Made progress but still blocked.
>
> After recompiling the code on cmd instead of PowerShell, now I can see all
> 5 classes as you mentioned.
>
>
>  However I am still seeing the same error as before. Anything else I can
> check for?
>
>
>
> *From:* Judy Nash [mailto:judyn...@exchange.microsoft.com
> <judyn...@exchange.microsoft.com>]
> *Sent:* Monday, November 24, 2014 11:50 PM
> *To:* Cheng Lian; u...@spark.incubator.apache.org
> *Subject:* RE: latest Spark 1.2 thrift server fail with
> NoClassDefFoundError on Guava
>
>
>
> This is what I got from jar tf:
>
> org/spark-project/guava/common/base/Preconditions.class
>
> org/spark-project/guava/common/math/MathPreconditions.class
>
> com/clearspring/analytics/util/Preconditions.class
>
> parquet/Preconditions.class
>
>
>
> I seem to have the line that reported missing, but I am missing this file:
>
> com/google/inject/internal/util/$Preconditions.class
>
>
>
> Any suggestion on how to fix this?
>
> Very much appreciate the help as I am very new to Spark and open source
> technologies.
>
>
>
> *From:* Cheng Lian [mailto:lian.cs....@gmail.com <lian.cs....@gmail.com>]
> *Sent:* Monday, November 24, 2014 8:24 PM
> *To:* Judy Nash; u...@spark.incubator.apache.org
> *Subject:* Re: latest Spark 1.2 thrift server fail with
> NoClassDefFoundError on Guava
>
>
>
> Hm, I tried exactly the same commit and the build command locally, but
> couldn’t reproduce this.
>
> Usually this kind of errors are caused by classpath misconfiguration.
> Could you please try this to ensure corresponding Guava classes are
> included in the assembly jar you built?
>
> jar tf 
> assembly/target/scala-2.10/spark-assembly-1.2.1-SNAPSHOT-hadoop2.4.0.jar | 
> grep Preconditions
>
> On my machine I got these lines (the first line is the one reported as
> missing in your case):
>
> org/spark-project/guava/common/base/Preconditions.class
>
> org/spark-project/guava/common/math/MathPreconditions.class
>
> com/clearspring/analytics/util/Preconditions.class
>
> parquet/Preconditions.class
>
> com/google/inject/internal/util/$Preconditions.class
>
> On 11/25/14 6:25 AM, Judy Nash wrote:
>
> Thank you Cheng for responding.
>
>
> Here is the commit SHA1 on the 1.2 branch I saw this failure in:
>
> commit 6f70e0295572e3037660004797040e026e440dbd
>
> Author: zsxwing <zsxw...@gmail.com> <zsxw...@gmail.com>
>
> Date:   Fri Nov 21 00:42:43 2014 -0800
>
>
>
>     [SPARK-4472][Shell] Print "Spark context available as sc." only when
> SparkContext is created...
>
>
>
>     ... successfully
>
>
>
>     It's weird that printing "Spark context available as sc" when creating
> SparkContext unsuccessfully.
>
>
>
> Let me know if you need anything else.
>
>
>
> *From:* Cheng Lian [mailto:lian.cs....@gmail.com <lian.cs....@gmail.com>]
> *Sent:* Friday, November 21, 2014 8:02 PM
> *To:* Judy Nash; u...@spark.incubator.apache.org
> *Subject:* Re: latest Spark 1.2 thrift server fail with
> NoClassDefFoundError on Guava
>
>
>
> Hi Judy, could you please provide the commit SHA1 of the version you're
> using? Thanks!
>
> On 11/22/14 11:05 AM, Judy Nash wrote:
>
> Hi,
>
>
>
> Thrift server is failing to start for me on latest spark 1.2 branch.
>
>
>
> I got the error below when I start thrift server.
>
> Exception in thread "main" java.lang.NoClassDefFoundError:
> com/google/common/bas
>
> e/Preconditions
>
>         at
> org.apache.hadoop.conf.Configuration$DeprecationDelta.<init>(Configur
>
> ation.java:314)….
>
>
>
> Here is my setup:
>
> 1)      Latest spark 1.2 branch build
>
> 2)      Used build command:
>
> mvn -Pyarn -Phadoop-2.4 -Dhadoop.version=2.4.0 -Phive -Phive-thriftserver
> -DskipTests clean package
>
> 3)      Added hive-site.xml to \conf
>
> 4)      Version on the box: Hive 0.13, Hadoop 2.4
>
>
>
> Is this a real bug or am I doing something wrong?
>
>
>
> -----------------------------------
>
> Full Stacktrace:
>
> Exception in thread "main" java.lang.NoClassDefFoundError:
> com/google/common/bas
>
> e/Preconditions
>
>         at
> org.apache.hadoop.conf.Configuration$DeprecationDelta.<init>(Configur
>
> ation.java:314)
>
>         at
> org.apache.hadoop.conf.Configuration$DeprecationDelta.<init>(Configur
>
> ation.java:327)
>
>         at
> org.apache.hadoop.conf.Configuration.<clinit>(Configuration.java:409)
>
>
>
>         at
> org.apache.spark.deploy.SparkHadoopUtil.newConfiguration(SparkHadoopU
>
> til.scala:82)
>
>         at
> org.apache.spark.deploy.SparkHadoopUtil.<init>(SparkHadoopUtil.scala:
>
> 42)
>
>         at
> org.apache.spark.deploy.SparkHadoopUtil$.<init>(SparkHadoopUtil.scala
>
> :202)
>
>         at
> org.apache.spark.deploy.SparkHadoopUtil$.<clinit>(SparkHadoopUtil.sca
>
> la)
>
>         at
> org.apache.spark.util.Utils$.getSparkOrYarnConfig(Utils.scala:1784)
>
>         at
> org.apache.spark.storage.BlockManager.<init>(BlockManager.scala:105)
>
>         at
> org.apache.spark.storage.BlockManager.<init>(BlockManager.scala:180)
>
>         at org.apache.spark.SparkEnv$.create(SparkEnv.scala:292)
>
>         at org.apache.spark.SparkEnv$.createDriverEnv(SparkEnv.scala:159)
>
>         at org.apache.spark.SparkContext.<init>(SparkContext.scala:230)
>
>         at
> org.apache.spark.sql.hive.thriftserver.SparkSQLEnv$.init(SparkSQLEnv.
>
> scala:38)
>
>         at
> org.apache.spark.sql.hive.thriftserver.HiveThriftServer2$.main(HiveTh
>
> riftServer2.scala:56)
>
>         at
> org.apache.spark.sql.hive.thriftserver.HiveThriftServer2.main(HiveThr
>
> iftServer2.scala)
>
>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>
>         at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.
>
> java:57)
>
>         at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAcces
>
> sorImpl.java:43)
>
>         at java.lang.reflect.Method.invoke(Method.java:606)
>
>         at
> org.apache.spark.deploy.SparkSubmit$.launch(SparkSubmit.scala:353)
>
>         at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:75)
>
>         at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
>
> Caused by: java.lang.ClassNotFoundException:
> com.google.common.base.Precondition
>
> s
>
>         at java.net.URLClassLoader$1.run(URLClassLoader.java:366)
>
>         at java.net.URLClassLoader$1.run(URLClassLoader.java:355)
>
>         at java.security.AccessController.doPrivileged(Native Method)
>
>         at java.net.URLClassLoader.findClass(URLClassLoader.java:354)
>
>         at java.lang.ClassLoader.loadClass(ClassLoader.java:425)
>
>         at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308)
>
>         at java.lang.ClassLoader.loadClass(ClassLoader.java:358)
>
>
>
>  ​
>
>
>

Reply via email to