Enthusiast sparkenthusi...@yahoo.in
wrote:
All examples of Spark Stream programming that I see assume streams of
lines that are then tokenised and acted upon (like the WordCount example).
How do I process Streams that span multiple lines? Are there examples that
I can use?
All examples of Spark Stream programming that I see assume streams of lines
that are then tokenised and acted upon (like the WordCount example).
How do I process Streams that span multiple lines? Are there examples that I
can use?
Are you looking for RDD.wholeTextFiles?
On 3 August 2015 at 10:57, Spark Enthusiast sparkenthusi...@yahoo.in
wrote:
All examples of Spark Stream programming that I see assume streams of
lines that are then tokenised and acted upon (like the WordCount example).
How do I process Streams
that I see assume streams of
lines that are then tokenised and acted upon (like the WordCount example).
How do I process Streams that span multiple lines? Are there examples
that I can use?