Matei, thank you. That seemed to work but I'm not able to import a class from my jar.
Using the verbose options, I can see that my jar should be included Parsed arguments: ... jars /Users/rhoover/Work/spark-etl/target/scala-2.10/spark-etl_2.10-1.0.jar And I see the class I want to load in the jar: jar -tf /Users/rhoover/Work/spark-etl/target/scala-2.10/spark-etl_2.10-1.0.jar | grep IP2IncomeJob etl/IP2IncomeJob$$anonfun$1.class etl/IP2IncomeJob$$anonfun$4.class etl/IP2IncomeJob$.class etl/IP2IncomeJob$$anonfun$splitOverlappingRange$1.class etl/IP2IncomeJob.class etl/IP2IncomeJob$$anonfun$3.class etl/IP2IncomeJob$$anonfun$2.class But the import fails scala> import etl.IP2IncomeJob <console>:10: error: not found: value etl import etl.IP2IncomeJob Any ideas? On Sun, Apr 27, 2014 at 3:46 PM, Matei Zaharia <matei.zaha...@gmail.com>wrote: > Hi Roger, > > You should be able to use the --jars argument of spark-shell to add JARs > onto the classpath and then work with those classes in the shell. (A recent > patch, https://github.com/apache/spark/pull/542, made spark-shell use the > same command-line arguments as spark-submit). But this is a great question, > we should test it out and see whether anything else would make development > easier. > > SBT also has an interactive shell where you can run classes in your > project, but unfortunately Spark can’t deal with closures typed directly in > that the right way. However you write your Spark logic in a method and just > call that method from the SBT shell, that should work. > > Matei > > On Apr 27, 2014, at 3:14 PM, Roger Hoover <roger.hoo...@gmail.com> wrote: > > > Hi, > > > > From the meetup talk about the 1.0 release, I saw that spark-submit will > be the preferred way to launch apps going forward. > > > > How do you recommend launching such jobs in a development cycle? For > example, how can I load an app that's expecting to a given to spark-submit > into spark-shell? > > > > Also, can anyone recommend other tricks for rapid development? I'm new > to Scala, sbt, etc. I think sbt can watch for changes in source files and > compile them automatically. > > > > I want to be able to make code changes and quickly get into a > spark-shell to play around with them. > > > > I appreciate any advice. Thanks, > > > > Roger > >