There is a way to :load in shell, where you can specify the path of your
boilerplate.scala. These things would be streamlined "once" we have scala
2.11 (I hope.)


On Tue, Feb 18, 2014 at 7:20 AM, Mayur Rustagi <[email protected]>wrote:

> That's actually not a bad idea. To have a shellboilerplate.scala in the
> same folder that is used to initialize the shell.
> Shell is a script that end of they day starts a JVM with jars from the
> spark project , mostly you'll have to modify the spark classes and
> reassemble using sbt. It's messy but thr may be easier ways to feed some
> data to shell script/JVM then connect with stdin.
> Regards
> Mayur
>
> On Monday, February 17, 2014, Kyle Ellrott <[email protected]> wrote:
>
>> Is there a way to define a set of commands to 'initialize' the
>> environment in the SparkShell?
>> I'd like to create a wrapper script that starts up the sparkshell and
>> does some boiler plate initialization (imports and variable
>> creation) before handing things over to me.
>>
>> Kyle
>>
>
>
> --
> Sent from Gmail Mobile
>



-- 
Prashant

Reply via email to