TaskNotSerializableException when running through Spark shell

2014-10-16 Thread Akshat Aranya
Hi, Can anyone explain how things get captured in a closure when runing through the REPL. For example: def foo(..) = { .. } rdd.map(foo) sometimes complains about classes not being serializable that are completely unrelated to foo. This happens even when I write it such: object Foo { def

Re: TaskNotSerializableException when running through Spark shell

2014-10-16 Thread Jimmy McErlain
I actually only ran into this issue recently after we upgraded to Spark 1.1. Within the REPL for Spark 1.0 everything works fine but within the REPL for 1.1, it is not. FYI I am also only doing simple regex matching functions within an RDD... Now when I am running the same code as App everything