Are you looking for RDD.wholeTextFiles? On 3 August 2015 at 10:57, Spark Enthusiast <sparkenthusi...@yahoo.in> wrote:
> All examples of Spark Stream programming that I see assume streams of > lines that are then tokenised and acted upon (like the WordCount example). > > How do I process Streams that span multiple lines? Are there examples that > I can use? >