Hi, I'm new to Scala but working on a simple patch for a configuration change and want to run just my unit tests. When I run ./sbt test-only, it executes all sorts of other tests but not the one I want. Is there an easy way to run a single test? Any help is appreciated.
$ ./sbt test-only kafka.utils.JsonTest [info] Loading project definition from /Users/rhoover/Work/kafka/project [warn] Multiple resolvers having different access mechanism configured with same name 'sbt-plugin-releases'. To avoid conflict, Remove duplicate project resolvers (`resolvers`) or rename publishing resolver (`publishTo`). [info] Set current project to Kafka (in build file:/Users/rhoover/Work/kafka/) [info] No tests to run for Kafka/test:test-only [info] No tests to run for contrib/test:test-only [info] No tests to run for java-examples/test:test-only [info] No tests to run for perf/test:test-only [info] No tests to run for hadoop-producer/test:test-only [info] No tests to run for hadoop-consumer/test:test-only [info] Test Starting: testFetcher(kafka.integration.FetcherTest) [info] Test Passed: testFetcher(kafka.integration.FetcherTest) [2013-10-25 16:19:55,067] ERROR Error in cleaner thread 0: (kafka.log.LogCleaner:103) java.lang.IllegalArgumentException: inconsistent range at java.util.concurrent.ConcurrentSkipListMap$SubMap.<init>(ConcurrentSkipListMap.java:2506) at java.util.concurrent.ConcurrentSkipListMap.subMap(ConcurrentSkipListMap.java:1984) at kafka.log.Log.logSegments(Log.scala:604) at kafka.log.LogToClean.<init>(LogCleaner.scala:596) at kafka.log.LogCleaner$$anonfun$5.apply(LogCleaner.scala:137) at kafka.log.LogCleaner$$anonfun$5.apply(LogCleaner.scala:137) at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206) at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206) at scala.collection.LinearSeqOptimized$class.foreach(LinearSeqOptimized.scala:61) at scala.collection.immutable.List.foreach(List.scala:45) at scala.collection.TraversableLike$class.map(TraversableLike.scala:206) at scala.collection.immutable.List.map(List.scala:45) at kafka.log.LogCleaner.kafka$log$LogCleaner$$grabFilthiestLog(LogCleaner.scala:137) at kafka.log.LogCleaner$CleanerThread.cleanOrSleep(LogCleaner.scala:203) at kafka.log.LogCleaner$CleanerThread.run(LogCleaner.scala:189) #THIS DOESN'T WORK EITHER BUT EXECUTES OTHER TESTS $ ./sbt test-only unit.kafka.utils.JsonTest [info] Loading project definition from /Users/rhoover/Work/kafka/project [warn] Multiple resolvers having different access mechanism configured with same name 'sbt-plugin-releases'. To avoid conflict, Remove duplicate project resolvers (`resolvers`) or rename publishing resolver (`publishTo`). [info] Set current project to Kafka (in build file:/Users/rhoover/Work/kafka/) [info] No tests to run for contrib/test:test-only [info] No tests to run for Kafka/test:test-only [info] No tests to run for java-examples/test:test-only [info] No tests to run for perf/test:test-only [info] No tests to run for hadoop-consumer/test:test-only [info] No tests to run for hadoop-producer/test:test-only [info] Test Starting: truncate [info] Test Passed: truncate [info] Test Starting: randomLookupTest [info] Test Passed: randomLookupTest [info] Test Starting: lookupExtremeCases [info] Test Passed: lookupExtremeCases [info] Test Starting: appendTooMany [info] Test Passed: appendTooMany [info] Test Starting: appendOutOfOrder [info] Test Passed: appendOutOfOrder [info] Test Starting: testReopen [info] Test Passed: testReopen [2013-10-25 16:21:44,004] ERROR [KafkaApi-0] Error when processing fetch request for partition [test2,0] offset -1 from consumer with correlation id 0 (kafka.server.KafkaApis:103) kafka.common.OffsetOutOfRangeException: Request for offset -1 but we only have log segments in the range 0 to 2. at kafka.log.Log.read(Log.scala:380) at kafka.server.KafkaApis.kafka$server$KafkaApis$$readMessageSet(KafkaApis.scala:401) at kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1.apply(KafkaApis.scala:347) at kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1.apply(KafkaApis.scala:343) at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206) at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206) at scala.collection.immutable.Map$Map4.foreach(Map.scala:180) at scala.collection.TraversableLike$class.map(TraversableLike.scala:206) at scala.collection.immutable.Map$Map4.map(Map.scala:157) at kafka.server.KafkaApis.kafka$server$KafkaApis$$readMessageSets(KafkaApis.scala:343) at kafka.server.KafkaApis.handleFetchRequest(KafkaApis.scala:309) at kafka.server.KafkaApis.handle(KafkaApis.scala:71) at kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:42) at java.lang.Thread.run(Thread.java:744)