Hi Janardhan, What's the command to build the project (sbt package or sbt assembly)? What's the command you execute to run the application?
Pozdrawiam, Jacek Laskowski ---- https://medium.com/@jaceklaskowski/ Mastering Apache Spark 2.0 http://bit.ly/mastering-apache-spark Follow me at https://twitter.com/jaceklaskowski On Mon, Sep 19, 2016 at 2:12 AM, janardhan shetty <janardhan...@gmail.com> wrote: > Hi Sujit, > > Tried that option but same error: > > java version "1.8.0_51" > > > libraryDependencies ++= { > val sparkVersion = "2.0.0" > Seq( > "org.apache.spark" %% "spark-core" % sparkVersion % "provided", > "org.apache.spark" %% "spark-sql" % sparkVersion % "provided", > "org.apache.spark" %% "spark-streaming" % sparkVersion % "provided", > "org.apache.spark" %% "spark-mllib" % sparkVersion % "provided", > "edu.stanford.nlp" % "stanford-corenlp" % "3.6.0", > "com.google.protobuf" % "protobuf-java" % "2.6.1", > "edu.stanford.nlp" % "stanford-corenlp" % "3.6.0", > "org.scalatest" %% "scalatest" % "2.2.6" % "test" > ) > } > > Error: > > Exception in thread "main" java.lang.NoClassDefFoundError: > edu/stanford/nlp/pipeline/StanfordCoreNLP > at > transformers.ml.Lemmatizer$$anonfun$createTransformFunc$1.apply(Lemmatizer.scala:37) > at > transformers.ml.Lemmatizer$$anonfun$createTransformFunc$1.apply(Lemmatizer.scala:33) > at > org.apache.spark.sql.catalyst.expressions.ScalaUDF$$anonfun$2.apply(ScalaUDF.scala:88) > at > org.apache.spark.sql.catalyst.expressions.ScalaUDF$$anonfun$2.apply(ScalaUDF.scala:87) > at > org.apache.spark.sql.catalyst.expressions.ScalaUDF.eval(ScalaUDF.scala:1060) > at > org.apache.spark.sql.catalyst.expressions.Alias.eval(namedExpressions.scala:142) > at > org.apache.spark.sql.catalyst.expressions.InterpretedProjection.apply(Projection.scala:45) > at > org.apache.spark.sql.catalyst.expressions.InterpretedProjection.apply(Projection.scala:29) > at > scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234) > at > scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234) > at scala.collection.immutable.List.foreach(List.scala:381) > at scala.collection.TraversableLike$class.map(TraversableLike.scala:234) > > > > On Sun, Sep 18, 2016 at 2:21 PM, Sujit Pal <sujitatgt...@gmail.com> wrote: >> >> Hi Janardhan, >> >> Maybe try removing the string "test" from this line in your build.sbt? >> IIRC, this restricts the models JAR to be called from a test. >> >> "edu.stanford.nlp" % "stanford-corenlp" % "3.6.0" % "test" classifier >> "models", >> >> -sujit >> >> >> On Sun, Sep 18, 2016 at 11:01 AM, janardhan shetty >> <janardhan...@gmail.com> wrote: >>> >>> Hi, >>> >>> I am trying to use lemmatization as a transformer and added belwo to the >>> build.sbt >>> >>> "edu.stanford.nlp" % "stanford-corenlp" % "3.6.0", >>> "com.google.protobuf" % "protobuf-java" % "2.6.1", >>> "edu.stanford.nlp" % "stanford-corenlp" % "3.6.0" % "test" classifier >>> "models", >>> "org.scalatest" %% "scalatest" % "2.2.6" % "test" >>> >>> >>> Error: >>> Exception in thread "main" java.lang.NoClassDefFoundError: >>> edu/stanford/nlp/pipeline/StanfordCoreNLP >>> >>> I have tried other versions of this spark package. >>> >>> Any help is appreciated.. >> >> > --------------------------------------------------------------------- To unsubscribe e-mail: user-unsubscr...@spark.apache.org