Hi all,

For each record I’m processing in a Spark streaming app (written in Java) I
need to take over 30 datapoints.

The output of my map would be something like:
KEY1,1,0,1,0,30,1,1,1,1,0,30,…
KEY1,0,1,1,0,15,1,1,1,1,0,28,…
KEY2,0,1,1,0,22,1,1,1,1,0,0,…

And I want to end up with:
KEY1,1,1,2,0,45,2,2,2,2,0,58,…
KEY2,0,1,1,0,22,1,1,1,1,0,0,…

What’s the best way to achieve that in Spark? Any problems with this
approach?

Cheers,
Kev

Reply via email to