Sounds reasonable.

Please consider posting question on Spark C* connector on their mailing
list if you have any.

On Sun, Feb 14, 2016 at 7:51 PM, Kevin Burton <bur...@spinn3r.com> wrote:

> Afternoon.
>
> About 6 months ago I tried (and failed) to get Spark and Cassandra working
> together in production due to dependency hell.
>
> I'm going to give it another try!
>
> Here's my general strategy.
>
> I'm going to create a maven module for my code... with spark dependencies.
>
> Then I'm going to get that to run and have unit tests for reading from
> files and writing the data back out the way I want via spark jobs.
>
> Then I'm going to setup cassandra unit to embed cassandra in my project.
> Then I'm going to point Spark to Cassandra and have the same above code
> work with Cassandra but instead of reading from a file it reads/writes to
> C*.
>
> Then once testing is working I'm going to setup spark in cluster mode with
> the same dependencies.
>
> Does this sound like a reasonable strategy?
>
> Kevin
>
> --
>
> We’re hiring if you know of any awesome Java Devops or Linux Operations
> Engineers!
>
> Founder/CEO Spinn3r.com
> Location: *San Francisco, CA*
> blog: http://burtonator.wordpress.com
> … or check out my Google+ profile
> <https://plus.google.com/102718274791889610666/posts>
>
>

Reply via email to