Hi all, For each record I’m processing in a Spark streaming app (written in Java) I need to take over 30 datapoints.
The output of my map would be something like: KEY1,1,0,1,0,30,1,1,1,1,0,30,… KEY1,0,1,1,0,15,1,1,1,1,0,28,… KEY2,0,1,1,0,22,1,1,1,1,0,0,… And I want to end up with: KEY1,1,1,2,0,45,2,2,2,2,0,58,… KEY2,0,1,1,0,22,1,1,1,1,0,0,… What’s the best way to achieve that in Spark? Any problems with this approach? Cheers, Kev