Same error. I saw someone reported the same issue, e.g.
http://apache-spark-user-list.1001560.n3.nabble.com/Spark-streaming-kafka-error-td9106.html

Should I use "sbt assembly"? It failed for deduplicate though.

error] (*:assembly) deduplicate: different file contents found in the
following:
[error]
/Users/gzhao/.ivy2/cache/org.eclipse.jetty.orbit/javax.transaction/orbits/javax.transaction-1.1.1.v201105210645.jar:META-INF/ECLIPSEF.RSA
[error]
/Users/gzhao/.ivy2/cache/org.eclipse.jetty.orbit/javax.servlet/orbits/javax.servlet-3.0.0.v201112011016.jar:META-INF/ECLIPSEF.RSA
[error]
/Users/gzhao/.ivy2/cache/org.eclipse.jetty.orbit/javax.mail.glassfish/orbits/javax.mail.glassfish-1.4.1.v201005082020.jar:META-INF/ECLIPSEF.RSA
[error]
/Users/gzhao/.ivy2/cache/org.eclipse.jetty.orbit/javax.activation/orbits/javax.activation-1.1.0.v201105071233.jar:META-INF/ECLIPSEF.RSA
[error] Total time: 4 s, completed Oct 16, 2014 1:58:41 PM


On Thu, Oct 16, 2014 at 12:11 PM, Akhil Das <ak...@sigmoidanalytics.com>
wrote:

> Can you try:
>
> sbt:
>
> name := "Simple Project"
>
>
> version := "1.1"
>
>
> scalaVersion := "2.10.4"
>
>
> libraryDependencies ++= Seq(
>
>     "org.apache.spark" %% "spark-core" % "1.1.0",
>
>     "org.apache.spark" %% "spark-streaming" % "1.1.0",
>
>     "org.apache.spark" %% "spark-streaming-kafka" % "1.1.0"
>
>   )
>
> Thanks
> Best Regards
>
> On Fri, Oct 17, 2014 at 12:36 AM, Gary Zhao <garyz...@gmail.com> wrote:
>
>> Thanks Akhil. I tried spark-submit and saw the same issue. I double
>> checked the versions and they look ok. Are you seeing any obvious issues?
>>
>> sbt:
>>
>> name := "Simple Project"
>>
>>
>> version := "1.1"
>>
>>
>> scalaVersion := "2.10.4"
>>
>>
>> libraryDependencies ++= Seq(
>>
>>     "org.apache.spark" %% "spark-core" % "1.1.0",
>>
>>     "org.apache.spark" %% "spark-streaming" % "1.1.0",
>>
>>     "org.apache.spark" %% "spark-streaming-kafka" % "1.1.0",
>>
>>     "org.apache.kafka" %% "kafka" % "0.8.0"
>>
>>   )
>>
>>
>> spark-1.1.0-bin-hadoop1/bin/spark-submit --class "main.scala.SimpleApp"
>> --master "local[2]" simple-project_2.10-1.1.jar --jars
>> spark-streaming-kafka_2.10-1.1.0.jar,kafka_2.10-0.8.0.jar
>>
>> Exception in thread "main" java.lang.NoClassDefFoundError:
>> org/apache/spark/streaming/kafka/KafkaUtils$
>> at main.scala.SimpleApp$delayedInit$body.apply(SimpleApp.scala:15)
>> at scala.Function0$class.apply$mcV$sp(Function0.scala:40)
>> at
>> scala.runtime.AbstractFunction0.apply$mcV$sp(AbstractFunction0.scala:12)
>> at scala.App$$anonfun$main$1.apply(App.scala:71)
>> at scala.App$$anonfun$main$1.apply(App.scala:71)
>> at scala.collection.immutable.List.foreach(List.scala:318)
>> at
>> scala.collection.generic.TraversableForwarder$class.foreach(TraversableForwarder.scala:32)
>> at scala.App$class.main(App.scala:71)
>> at main.scala.SimpleApp$.main(SimpleApp.scala:11)
>> at main.scala.SimpleApp.main(SimpleApp.scala)
>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>> at
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>> at
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>> at java.lang.reflect.Method.invoke(Method.java:606)
>> at org.apache.spark.deploy.SparkSubmit$.launch(SparkSubmit.scala:328)
>> at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:75)
>> at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
>> Caused by: java.lang.ClassNotFoundException:
>> org.apache.spark.streaming.kafka.KafkaUtils$
>> at java.net.URLClassLoader$1.run(URLClassLoader.java:366)
>> at java.net.URLClassLoader$1.run(URLClassLoader.java:355)
>> at java.security.AccessController.doPrivileged(Native Method)
>> at java.net.URLClassLoader.findClass(URLClassLoader.java:354)
>> at java.lang.ClassLoader.loadClass(ClassLoader.java:424)
>> at java.lang.ClassLoader.loadClass(ClassLoader.java:357)
>> ... 17 more
>>
>>
>> On Tue, Oct 14, 2014 at 12:05 AM, Akhil Das <ak...@sigmoidanalytics.com>
>> wrote:
>>
>>> Just make sure you have the same version of spark-streaming-kafka
>>> <http://mvnrepository.com/artifact/org.apache.spark/spark-streaming-kafka_2.10>
>>> jar and spark in your classpath.
>>>
>>> Thanks
>>> Best Regards
>>>
>>> On Tue, Oct 14, 2014 at 9:02 AM, Gary Zhao <garyz...@gmail.com> wrote:
>>>
>>>> Hello
>>>>
>>>> I'm trying to connect kafka in spark shell, but failed as below. Could
>>>> you take a look what I missed.
>>>>
>>>> scala>  val kafkaStream = KafkaUtils.createStream(ssc,
>>>> "test-vip.snc1:2181", "test_spark", Map("user-test"->1))
>>>> error: bad symbolic reference. A signature in KafkaUtils.class refers
>>>> to term serializer
>>>> in value kafka which is not available.
>>>> It may be completely missing from the current classpath, or the version
>>>> on
>>>> the classpath might be incompatible with the version used when
>>>> compiling KafkaUtils.class.
>>>>
>>>> Thanks
>>>> Gary
>>>>
>>>
>>>
>>
>

Reply via email to