Thanks, Andrew.  I'll give it a try.

On Mon, May 26, 2014 at 2:22 PM, Andrew Or <and...@databricks.com> wrote:

> Hi Roger,
>
> This was due to a bug in the Spark shell code, and is fixed in the latest
> master (and RC11). Here is the commit that fixed it:
> https://github.com/apache/spark/commit/8edbee7d1b4afc192d97ba192a5526affc464205.
> Try it now and it should work. :)
>
> Andrew
>
>
> 2014-05-26 10:35 GMT+02:00 Perttu Ranta-aho <ranta...@iki.fi>:
>
> Hi Roger,
>>
>> Were you able to solve this?
>>
>> -Perttu
>>
>>
>> On Tue, Apr 29, 2014 at 8:11 AM, Roger Hoover <roger.hoo...@gmail.com>wrote:
>>
>>> Patrick,
>>>
>>> Thank you for replying.  That didn't seem to work either.  I see the
>>> option parsed using verbose mode.
>>>
>>> Parsed arguments:
>>>  ...
>>>   driverExtraClassPath
>>>  /Users/rhoover/Work/spark-etl/target/scala-2.10/spark-etl_2.10-1.0.jar
>>>
>>> But the jar still doesn't show up if I run ":cp" in the repl and the
>>> import still fails.
>>>
>>> scala> import etl._
>>> <console>:7: error: not found: value etl
>>>        import etl._
>>>
>>> Not sure if this helps, but I noticed with Spark 0.9.1 that the import
>>> only seems to work went I add the -usejavacp option to the spark-shell
>>> command.  I don't really understand why.
>>>
>>> With the latest code, I tried adding these options to the spark-shell
>>> command without success: -usejavacp -Dscala.usejavacp=true
>>>
>>>
>>> On Mon, Apr 28, 2014 at 6:30 PM, Patrick Wendell <pwend...@gmail.com>wrote:
>>>
>>>> What about if you run ./bin/spark-shell
>>>> --driver-class-path=/path/to/your/jar.jar
>>>>
>>>> I think either this or the --jars flag should work, but it's possible
>>>> there is a bug with the --jars flag when calling the Repl.
>>>>
>>>>
>>>> On Mon, Apr 28, 2014 at 4:30 PM, Roger Hoover 
>>>> <roger.hoo...@gmail.com>wrote:
>>>>
>>>>> A couple of issues:
>>>>> 1) the jar doesn't show up on the classpath even though SparkSubmit
>>>>> had it in the --jars options.  I tested this by running > :cp in 
>>>>> spark-shell
>>>>> 2) After adding it the classpath using (:cp
>>>>> /Users/rhoover/Work/spark-etl/target/scala-2.10/spark-etl_2.10-1.0.jar), 
>>>>> it
>>>>> still fails.  When I do that in the scala repl, it works.
>>>>>
>>>>> BTW, I'm using the latest code from the master branch
>>>>> (8421034e793c0960373a0a1d694ce334ad36e747)
>>>>>
>>>>>
>>>>> On Mon, Apr 28, 2014 at 3:40 PM, Roger Hoover 
>>>>> <roger.hoo...@gmail.com>wrote:
>>>>>
>>>>>> Matei,  thank you.  That seemed to work but I'm not able to import a
>>>>>> class from my jar.
>>>>>>
>>>>>> Using the verbose options, I can see that my jar should be included
>>>>>>
>>>>>> Parsed arguments:
>>>>>> ...
>>>>>>   jars
>>>>>>  /Users/rhoover/Work/spark-etl/target/scala-2.10/spark-etl_2.10-1.0.jar
>>>>>>
>>>>>> And I see the class I want to load in the jar:
>>>>>>
>>>>>> jar -tf
>>>>>> /Users/rhoover/Work/spark-etl/target/scala-2.10/spark-etl_2.10-1.0.jar |
>>>>>> grep IP2IncomeJob
>>>>>> etl/IP2IncomeJob$$anonfun$1.class
>>>>>> etl/IP2IncomeJob$$anonfun$4.class
>>>>>> etl/IP2IncomeJob$.class
>>>>>> etl/IP2IncomeJob$$anonfun$splitOverlappingRange$1.class
>>>>>> etl/IP2IncomeJob.class
>>>>>> etl/IP2IncomeJob$$anonfun$3.class
>>>>>> etl/IP2IncomeJob$$anonfun$2.class
>>>>>>
>>>>>> But the import fails
>>>>>>
>>>>>> scala> import etl.IP2IncomeJob
>>>>>> <console>:10: error: not found: value etl
>>>>>>        import etl.IP2IncomeJob
>>>>>>
>>>>>> Any ideas?
>>>>>>
>>>>>>
>>>>>>
>>>>>> On Sun, Apr 27, 2014 at 3:46 PM, Matei Zaharia <
>>>>>> matei.zaha...@gmail.com> wrote:
>>>>>>
>>>>>>> Hi Roger,
>>>>>>>
>>>>>>> You should be able to use the --jars argument of spark-shell to add
>>>>>>> JARs onto the classpath and then work with those classes in the shell. 
>>>>>>> (A
>>>>>>> recent patch, https://github.com/apache/spark/pull/542, made
>>>>>>> spark-shell use the same command-line arguments as spark-submit). But 
>>>>>>> this
>>>>>>> is a great question, we should test it out and see whether anything else
>>>>>>> would make development easier.
>>>>>>>
>>>>>>> SBT also has an interactive shell where you can run classes in your
>>>>>>> project, but unfortunately Spark can’t deal with closures typed 
>>>>>>> directly in
>>>>>>> that the right way. However you write your Spark logic in a method and 
>>>>>>> just
>>>>>>> call that method from the SBT shell, that should work.
>>>>>>>
>>>>>>> Matei
>>>>>>>
>>>>>>> On Apr 27, 2014, at 3:14 PM, Roger Hoover <roger.hoo...@gmail.com>
>>>>>>> wrote:
>>>>>>>
>>>>>>> > Hi,
>>>>>>> >
>>>>>>> > From the meetup talk about the 1.0 release, I saw that
>>>>>>> spark-submit will be the preferred way to launch apps going forward.
>>>>>>> >
>>>>>>> > How do you recommend launching such jobs in a development cycle?
>>>>>>>  For example, how can I load an app that's expecting to a given to
>>>>>>> spark-submit into spark-shell?
>>>>>>> >
>>>>>>> > Also, can anyone recommend other tricks for rapid development?
>>>>>>>  I'm new to Scala, sbt, etc.  I think sbt can watch for changes in 
>>>>>>> source
>>>>>>> files and compile them automatically.
>>>>>>> >
>>>>>>> > I want to be able to make code changes and quickly get into a
>>>>>>> spark-shell to play around with them.
>>>>>>> >
>>>>>>> > I appreciate any advice.  Thanks,
>>>>>>> >
>>>>>>> > Roger
>>>>>>>
>>>>>>>
>>>>>>
>>>>>
>>>>
>>>
>>
>

Reply via email to