Hi,

I'm new to Scala but working on a simple patch for a configuration change
and want to run just my unit tests.  When I run ./sbt test-only, it
executes all sorts of other tests but not the one I want.  Is there an easy
way to run a single test?  Any help is appreciated.

$ ./sbt test-only kafka.utils.JsonTest
[info] Loading project definition from /Users/rhoover/Work/kafka/project
[warn] Multiple resolvers having different access mechanism configured with
same name 'sbt-plugin-releases'. To avoid conflict, Remove duplicate
project resolvers (`resolvers`) or rename publishing resolver (`publishTo`).
[info] Set current project to Kafka (in build
file:/Users/rhoover/Work/kafka/)
[info] No tests to run for Kafka/test:test-only
[info] No tests to run for contrib/test:test-only
[info] No tests to run for java-examples/test:test-only
[info] No tests to run for perf/test:test-only
[info] No tests to run for hadoop-producer/test:test-only
[info] No tests to run for hadoop-consumer/test:test-only
[info] Test Starting: testFetcher(kafka.integration.FetcherTest)
[info] Test Passed: testFetcher(kafka.integration.FetcherTest)
[2013-10-25 16:19:55,067] ERROR Error in cleaner thread 0:
(kafka.log.LogCleaner:103)
java.lang.IllegalArgumentException: inconsistent range
 at
java.util.concurrent.ConcurrentSkipListMap$SubMap.<init>(ConcurrentSkipListMap.java:2506)
at
java.util.concurrent.ConcurrentSkipListMap.subMap(ConcurrentSkipListMap.java:1984)
 at kafka.log.Log.logSegments(Log.scala:604)
at kafka.log.LogToClean.<init>(LogCleaner.scala:596)
 at kafka.log.LogCleaner$$anonfun$5.apply(LogCleaner.scala:137)
at kafka.log.LogCleaner$$anonfun$5.apply(LogCleaner.scala:137)
 at
scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
at
scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
 at
scala.collection.LinearSeqOptimized$class.foreach(LinearSeqOptimized.scala:61)
at scala.collection.immutable.List.foreach(List.scala:45)
 at scala.collection.TraversableLike$class.map(TraversableLike.scala:206)
at scala.collection.immutable.List.map(List.scala:45)
 at
kafka.log.LogCleaner.kafka$log$LogCleaner$$grabFilthiestLog(LogCleaner.scala:137)
at kafka.log.LogCleaner$CleanerThread.cleanOrSleep(LogCleaner.scala:203)
 at kafka.log.LogCleaner$CleanerThread.run(LogCleaner.scala:189)


#THIS DOESN'T WORK EITHER BUT EXECUTES OTHER TESTS

$ ./sbt test-only unit.kafka.utils.JsonTest
[info] Loading project definition from /Users/rhoover/Work/kafka/project
[warn] Multiple resolvers having different access mechanism configured with
same name 'sbt-plugin-releases'. To avoid conflict, Remove duplicate
project resolvers (`resolvers`) or rename publishing resolver (`publishTo`).
[info] Set current project to Kafka (in build
file:/Users/rhoover/Work/kafka/)
[info] No tests to run for contrib/test:test-only
[info] No tests to run for Kafka/test:test-only
[info] No tests to run for java-examples/test:test-only
[info] No tests to run for perf/test:test-only
[info] No tests to run for hadoop-consumer/test:test-only
[info] No tests to run for hadoop-producer/test:test-only
[info] Test Starting: truncate
[info] Test Passed: truncate
[info] Test Starting: randomLookupTest
[info] Test Passed: randomLookupTest
[info] Test Starting: lookupExtremeCases
[info] Test Passed: lookupExtremeCases
[info] Test Starting: appendTooMany
[info] Test Passed: appendTooMany
[info] Test Starting: appendOutOfOrder
[info] Test Passed: appendOutOfOrder
[info] Test Starting: testReopen
[info] Test Passed: testReopen
[2013-10-25 16:21:44,004] ERROR [KafkaApi-0] Error when processing fetch
request for partition [test2,0] offset -1 from consumer with correlation id
0 (kafka.server.KafkaApis:103)
kafka.common.OffsetOutOfRangeException: Request for offset -1 but we only
have log segments in the range 0 to 2.
at kafka.log.Log.read(Log.scala:380)
 at
kafka.server.KafkaApis.kafka$server$KafkaApis$$readMessageSet(KafkaApis.scala:401)
at
kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1.apply(KafkaApis.scala:347)
 at
kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1.apply(KafkaApis.scala:343)
at
scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
 at
scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
at scala.collection.immutable.Map$Map4.foreach(Map.scala:180)
 at scala.collection.TraversableLike$class.map(TraversableLike.scala:206)
at scala.collection.immutable.Map$Map4.map(Map.scala:157)
 at
kafka.server.KafkaApis.kafka$server$KafkaApis$$readMessageSets(KafkaApis.scala:343)
at kafka.server.KafkaApis.handleFetchRequest(KafkaApis.scala:309)
 at kafka.server.KafkaApis.handle(KafkaApis.scala:71)
at kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:42)
 at java.lang.Thread.run(Thread.java:744)

Reply via email to