Sorry, I meant any *other* SBT files.

However, what happens if you remove the line:

        exclude("org.eclipse.jetty.orbit", "javax.servlet")


dean


Dean Wampler, Ph.D.
Author: Programming Scala, 2nd Edition
<http://shop.oreilly.com/product/0636920033073.do> (O'Reilly)
Typesafe <http://typesafe.com>
@deanwampler <http://twitter.com/deanwampler>
http://polyglotprogramming.com

On Sun, Sep 14, 2014 at 11:53 AM, Dean Wampler <deanwamp...@gmail.com>
wrote:

> Can you post your whole SBT build file(s)?
>
> Dean Wampler, Ph.D.
> Author: Programming Scala, 2nd Edition
> <http://shop.oreilly.com/product/0636920033073.do> (O'Reilly)
> Typesafe <http://typesafe.com>
> @deanwampler <http://twitter.com/deanwampler>
> http://polyglotprogramming.com
>
> On Wed, Sep 10, 2014 at 6:48 AM, Thorsten Bergler <sp...@tbonline.de>
> wrote:
>
>> Hi,
>>
>> I just called:
>>
>> > test
>>
>> or
>>
>> > run
>>
>> Thorsten
>>
>>
>> Am 10.09.2014 um 13:38 schrieb arthur.hk.c...@gmail.com:
>>
>>  Hi,
>>>
>>> What is your SBT command and the parameters?
>>>
>>> Arthur
>>>
>>>
>>> On 10 Sep, 2014, at 6:46 pm, Thorsten Bergler <sp...@tbonline.de> wrote:
>>>
>>>  Hello,
>>>>
>>>> I am writing a Spark App which is already working so far.
>>>> Now I started to build also some UnitTests, but I am running into some
>>>> dependecy problems and I cannot find a solution right now. Perhaps someone
>>>> could help me.
>>>>
>>>> I build my Spark Project with SBT and it seems to be configured well,
>>>> because compiling, assembling and running the built jar with spark-submit
>>>> are working well.
>>>>
>>>> Now I started with the UnitTests, which I located under /src/test/scala.
>>>>
>>>> When I call "test" in sbt, I get the following:
>>>>
>>>> 14/09/10 12:22:06 INFO storage.BlockManagerMaster: Registered
>>>> BlockManager
>>>> 14/09/10 12:22:06 INFO spark.HttpServer: Starting HTTP Server
>>>> [trace] Stack trace suppressed: run last test:test for the full output.
>>>> [error] Could not run test test.scala.SetSuite: 
>>>> java.lang.NoClassDefFoundError:
>>>> javax/servlet/http/HttpServletResponse
>>>> [info] Run completed in 626 milliseconds.
>>>> [info] Total number of tests run: 0
>>>> [info] Suites: completed 0, aborted 0
>>>> [info] Tests: succeeded 0, failed 0, canceled 0, ignored 0, pending 0
>>>> [info] All tests passed.
>>>> [error] Error during tests:
>>>> [error]     test.scala.SetSuite
>>>> [error] (test:test) sbt.TestsFailedException: Tests unsuccessful
>>>> [error] Total time: 3 s, completed 10.09.2014 12:22:06
>>>>
>>>> last test:test gives me the following:
>>>>
>>>>  last test:test
>>>>>
>>>> [debug] Running TaskDef(test.scala.SetSuite,
>>>> org.scalatest.tools.Framework$$anon$1@6e5626c8, false, [SuiteSelector])
>>>> java.lang.NoClassDefFoundError: javax/servlet/http/HttpServletResponse
>>>>     at org.apache.spark.HttpServer.start(HttpServer.scala:54)
>>>>     at org.apache.spark.broadcast.HttpBroadcast$.createServer(
>>>> HttpBroadcast.scala:156)
>>>>     at org.apache.spark.broadcast.HttpBroadcast$.initialize(
>>>> HttpBroadcast.scala:127)
>>>>     at org.apache.spark.broadcast.HttpBroadcastFactory.initialize(
>>>> HttpBroadcastFactory.scala:31)
>>>>     at org.apache.spark.broadcast.BroadcastManager.initialize(
>>>> BroadcastManager.scala:48)
>>>>     at org.apache.spark.broadcast.BroadcastManager.<init>(
>>>> BroadcastManager.scala:35)
>>>>     at org.apache.spark.SparkEnv$.create(SparkEnv.scala:218)
>>>>     at org.apache.spark.SparkContext.<init>(SparkContext.scala:202)
>>>>     at test.scala.SetSuite.<init>(SparkTest.scala:16)
>>>>
>>>> I also noticed right now, that sbt run is also not working:
>>>>
>>>> 14/09/10 12:44:46 INFO spark.HttpServer: Starting HTTP Server
>>>> [error] (run-main-2) java.lang.NoClassDefFoundError:
>>>> javax/servlet/http/HttpServletResponse
>>>> java.lang.NoClassDefFoundError: javax/servlet/http/HttpServletResponse
>>>>     at org.apache.spark.HttpServer.start(HttpServer.scala:54)
>>>>     at org.apache.spark.broadcast.HttpBroadcast$.createServer(
>>>> HttpBroadcast.scala:156)
>>>>     at org.apache.spark.broadcast.HttpBroadcast$.initialize(
>>>> HttpBroadcast.scala:127)
>>>>     at org.apache.spark.broadcast.HttpBroadcastFactory.initialize(
>>>> HttpBroadcastFactory.scala:31)
>>>>     at org.apache.spark.broadcast.BroadcastManager.initialize(
>>>> BroadcastManager.scala:48)
>>>>     at org.apache.spark.broadcast.BroadcastManager.<init>(
>>>> BroadcastManager.scala:35)
>>>>     at org.apache.spark.SparkEnv$.create(SparkEnv.scala:218)
>>>>     at org.apache.spark.SparkContext.<init>(SparkContext.scala:202)
>>>>     at main.scala.PartialDuplicateScanner$.main(
>>>> PartialDuplicateScanner.scala:29)
>>>>     at main.scala.PartialDuplicateScanner.main(
>>>> PartialDuplicateScanner.scala)
>>>>
>>>> Here is my Testprojekt.sbt file:
>>>>
>>>> name := "Testprojekt"
>>>>
>>>> version := "1.0"
>>>>
>>>> scalaVersion := "2.10.4"
>>>>
>>>> libraryDependencies ++= {
>>>>   Seq(
>>>>     "org.apache.lucene" % "lucene-core" % "4.9.0",
>>>>     "org.apache.lucene" % "lucene-analyzers-common" % "4.9.0",
>>>>     "org.apache.lucene" % "lucene-queryparser" % "4.9.0",
>>>>     ("org.apache.spark" %% "spark-core" % "1.0.2").
>>>>         exclude("org.mortbay.jetty", "servlet-api").
>>>>         exclude("commons-beanutils", "commons-beanutils-core").
>>>>         exclude("commons-collections", "commons-collections").
>>>>         exclude("commons-collections", "commons-collections").
>>>>         exclude("com.esotericsoftware.minlog", "minlog").
>>>>         exclude("org.eclipse.jetty.orbit", "javax.mail.glassfish").
>>>>         exclude("org.eclipse.jetty.orbit", "javax.transaction").
>>>>         exclude("org.eclipse.jetty.orbit", "javax.servlet")
>>>>   )
>>>> }
>>>>
>>>> resolvers += "Akka Repository" at "http://repo.akka.io/releases/";
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>> ---------------------------------------------------------------------
>>>> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
>>>> For additional commands, e-mail: user-h...@spark.apache.org
>>>>
>>>>
>>> ---------------------------------------------------------------------
>>> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
>>> For additional commands, e-mail: user-h...@spark.apache.org
>>>
>>>
>>
>> ---------------------------------------------------------------------
>> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
>> For additional commands, e-mail: user-h...@spark.apache.org
>>
>>
>

Reply via email to