try sbt clean command before build the app.

or delete .ivy2 ans .sbt  folders(not a good methode). Then try to rebuild
the project.


On Thu, Jun 5, 2014 at 11:45 AM, Sean Owen <so...@cloudera.com> wrote:

> I think this is SPARK-1949 again: https://github.com/apache/spark/pull/906
> I think this change fixed this issue for a few people using the SBT
> build, worth committing?
>
> On Thu, Jun 5, 2014 at 6:40 AM, Shrikar archak <shrika...@gmail.com>
> wrote:
> > Hi All,
> > Now that the Spark Version 1.0.0 is release there should not be any
> problem
> > with the local jars.
> > Shrikars-MacBook-Pro:SimpleJob shrikar$ cat simple.sbt
> > name := "Simple Project"
> >
> > version := "1.0"
> >
> > scalaVersion := "2.10.4"
> >
> > libraryDependencies ++= Seq("org.apache.spark" %% "spark-core" % "1.0.0",
> >                             "org.apache.spark" %% "spark-streaming" %
> > "1.0.0")
> >
> > resolvers += "Akka Repository" at "http://repo.akka.io/releases/";
> >
> > I am still having this issue
> > [error] (run-main) java.lang.NoClassDefFoundError:
> > javax/servlet/http/HttpServletResponse
> > java.lang.NoClassDefFoundError: javax/servlet/http/HttpServletResponse
> > at org.apache.spark.HttpServer.start(HttpServer.scala:54)
> > at
> >
> org.apache.spark.broadcast.HttpBroadcast$.createServer(HttpBroadcast.scala:156)
> > at
> >
> org.apache.spark.broadcast.HttpBroadcast$.initialize(HttpBroadcast.scala:127)
> > at
> >
> org.apache.spark.broadcast.HttpBroadcastFactory.initialize(HttpBroadcastFactory.scala:31)
> > at
> >
> org.apache.spark.broadcast.BroadcastManager.initialize(BroadcastManager.scala:48)
> > at
> >
> org.apache.spark.broadcast.BroadcastManager.<init>(BroadcastManager.scala:35)
> > at org.apache.spark.SparkEnv$.create(SparkEnv.scala:218)
> > at org.apache.spark.SparkContext.<init>(SparkContext.scala:202)
> >
> > Any help would be greatly appreciated.
> >
> > Thanks,
> > Shrikar
> >
> >
> > On Fri, May 23, 2014 at 3:58 PM, Shrikar archak <shrika...@gmail.com>
> wrote:
> >>
> >> Still the same error no change
> >>
> >> Thanks,
> >> Shrikar
> >>
> >>
> >> On Fri, May 23, 2014 at 2:38 PM, Jacek Laskowski <ja...@japila.pl>
> wrote:
> >>>
> >>> Hi Shrikar,
> >>>
> >>> How did you build Spark 1.0.0-SNAPSHOT on your machine? My
> >>> understanding is that `sbt publishLocal` is not enough and you really
> >>> need `sbt assembly` instead. Give it a try and report back.
> >>>
> >>> As to your build.sbt, upgrade Scala to 2.10.4 and "org.apache.spark"
> >>> %% "spark-streaming" % "1.0.0-SNAPSHOT" only that will pull down
> >>> spark-core as a transitive dep. The resolver for Akka Repository is
> >>> not needed. Your build.sbt should really look as follows:
> >>>
> >>> name := "Simple Project"
> >>>
> >>> version := "1.0"
> >>>
> >>> scalaVersion := "2.10.4"
> >>>
> >>> libraryDependencies += "org.apache.spark" %% "spark-streaming" %
> >>> "1.0.0-SNAPSHOT"
> >>>
> >>> Jacek
> >>>
> >>> On Thu, May 22, 2014 at 11:27 PM, Shrikar archak <shrika...@gmail.com>
> >>> wrote:
> >>> > Hi All,
> >>> >
> >>> > I am trying to run the network count example as a seperate standalone
> >>> > job
> >>> > and running into some issues.
> >>> >
> >>> > Environment:
> >>> > 1) Mac Mavericks
> >>> > 2) Latest spark repo from Github.
> >>> >
> >>> >
> >>> > I have a structure like this
> >>> >
> >>> > Shrikars-MacBook-Pro:SimpleJob shrikar$ find .
> >>> > .
> >>> > ./simple.sbt
> >>> > ./src
> >>> > ./src/main
> >>> > ./src/main/scala
> >>> > ./src/main/scala/NetworkWordCount.scala
> >>> > ./src/main/scala/SimpleApp.scala.bk
> >>> >
> >>> >
> >>> > simple.sbt
> >>> > name := "Simple Project"
> >>> >
> >>> > version := "1.0"
> >>> >
> >>> > scalaVersion := "2.10.3"
> >>> >
> >>> > libraryDependencies ++= Seq("org.apache.spark" %% "spark-core" %
> >>> > "1.0.0-SNAPSHOT",
> >>> >                             "org.apache.spark" %% "spark-streaming" %
> >>> > "1.0.0-SNAPSHOT")
> >>> >
> >>> > resolvers += "Akka Repository" at "http://repo.akka.io/releases/";
> >>> >
> >>> >
> >>> > I am able to run the SimpleApp which is mentioned in the doc but
> when I
> >>> > try
> >>> > to run the NetworkWordCount app I get error like this am I missing
> >>> > something?
> >>> >
> >>> > [info] Running com.shrikar.sparkapps.NetworkWordCount
> >>> > 14/05/22 14:26:47 INFO spark.SecurityManager: Changing view acls to:
> >>> > shrikar
> >>> > 14/05/22 14:26:47 INFO spark.SecurityManager: SecurityManager:
> >>> > authentication disabled; ui acls disabled; users with view
> permissions:
> >>> > Set(shrikar)
> >>> > 14/05/22 14:26:48 INFO slf4j.Slf4jLogger: Slf4jLogger started
> >>> > 14/05/22 14:26:48 INFO Remoting: Starting remoting
> >>> > 14/05/22 14:26:48 INFO Remoting: Remoting started; listening on
> >>> > addresses
> >>> > :[akka.tcp://spark@192.168.10.88:49963]
> >>> > 14/05/22 14:26:48 INFO Remoting: Remoting now listens on addresses:
> >>> > [akka.tcp://spark@192.168.10.88:49963]
> >>> > 14/05/22 14:26:48 INFO spark.SparkEnv: Registering MapOutputTracker
> >>> > 14/05/22 14:26:48 INFO spark.SparkEnv: Registering BlockManagerMaster
> >>> > 14/05/22 14:26:48 INFO storage.DiskBlockManager: Created local
> >>> > directory at
> >>> >
> >>> >
> /var/folders/r2/mbj08pb55n5d_9p8588xk5b00000gn/T/spark-local-20140522142648-0a14
> >>> > 14/05/22 14:26:48 INFO storage.MemoryStore: MemoryStore started with
> >>> > capacity 911.6 MB.
> >>> > 14/05/22 14:26:48 INFO network.ConnectionManager: Bound socket to
> port
> >>> > 49964
> >>> > with id = ConnectionManagerId(192.168.10.88,49964)
> >>> > 14/05/22 14:26:48 INFO storage.BlockManagerMaster: Trying to register
> >>> > BlockManager
> >>> > 14/05/22 14:26:48 INFO storage.BlockManagerInfo: Registering block
> >>> > manager
> >>> > 192.168.10.88:49964 with 911.6 MB RAM
> >>> > 14/05/22 14:26:48 INFO storage.BlockManagerMaster: Registered
> >>> > BlockManager
> >>> > 14/05/22 14:26:48 INFO spark.HttpServer: Starting HTTP Server
> >>> > [error] (run-main) java.lang.NoClassDefFoundError:
> >>> > javax/servlet/http/HttpServletResponse
> >>> > java.lang.NoClassDefFoundError:
> javax/servlet/http/HttpServletResponse
> >>> > at org.apache.spark.HttpServer.start(HttpServer.scala:54)
> >>> > at
> >>> >
> >>> >
> org.apache.spark.broadcast.HttpBroadcast$.createServer(HttpBroadcast.scala:156)
> >>> > at
> >>> >
> >>> >
> org.apache.spark.broadcast.HttpBroadcast$.initialize(HttpBroadcast.scala:127)
> >>> > at
> >>> >
> >>> >
> org.apache.spark.broadcast.HttpBroadcastFactory.initialize(HttpBroadcastFactory.scala:31)
> >>> > at
> >>> >
> >>> >
> org.apache.spark.broadcast.BroadcastManager.initialize(BroadcastManager.scala:48)
> >>> > at
> >>> >
> >>> >
> org.apache.spark.broadcast.BroadcastManager.<init>(BroadcastManager.scala:35)
> >>> > at org.apache.spark.SparkEnv$.create(SparkEnv.scala:218)
> >>> > at org.apache.spark.SparkContext.<init>(SparkContext.scala:202)
> >>> > at
> >>> >
> >>> >
> org.apache.spark.streaming.StreamingContext$.createNewSparkContext(StreamingContext.scala:549)
> >>> > at
> >>> >
> >>> >
> org.apache.spark.streaming.StreamingContext$.createNewSparkContext(StreamingContext.scala:561)
> >>> > at
> >>> >
> >>> >
> org.apache.spark.streaming.StreamingContext.<init>(StreamingContext.scala:91)
> >>> > at
> >>> >
> com.shrikar.sparkapps.NetworkWordCount$.main(NetworkWordCount.scala:39)
> >>> > at
> com.shrikar.sparkapps.NetworkWordCount.main(NetworkWordCount.scala)
> >>> > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> >>> > at
> >>> >
> >>> >
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
> >>> > at
> >>> >
> >>> >
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
> >>> > at java.lang.reflect.Method.invoke(Method.java:597)
> >>> >
> >>> >
> >>> > Thanks,
> >>> > Shrikar
> >>> >
> >>>
> >>>
> >>>
> >>> --
> >>> Jacek Laskowski | http://blog.japila.pl
> >>> "Never discourage anyone who continually makes progress, no matter how
> >>> slow." Plato
> >>
> >>
> >
>

Reply via email to