You can see an example of the constructor for the class which executes a
job in my opening post.

I'm attempting to instantiate and run the class using the code below:

```
    val conf = new SparkConf()
      .setAppName(appNameBase.format("Test"))

    val connector = CassandraConnector(conf)

    val sc = new SparkContext(conf)

    // Set up the threadpool for running Jobs.
    val pool = Executors.newFixedThreadPool(10)

    pool.execute(new SecondRollup(sc, connector, start))
```

There is some surrounding code that then waits for all the jobs entered
into the thread pool to complete, although it's not really required at the
moment as I am only submitting one job until I get this issue straightened
out :)

Thanks,

Lee

On Fri, Jun 5, 2015 at 11:50 AM Marcelo Vanzin <van...@cloudera.com> wrote:

> On Fri, Jun 5, 2015 at 11:48 AM, Lee McFadden <splee...@gmail.com> wrote:
>
>> Initially I had issues passing the SparkContext to other threads as it is
>> not serializable.  Eventually I found that adding the @transient annotation
>> prevents a NotSerializableException.
>>
>
> This is really puzzling. How are you passing the context around that you
> need to do serialization?
>
> Threads run all in the same process so serialization should not be needed
> at all.
>
> --
> Marcelo
>

Reply via email to