>From >examples/src/main/scala/org/apache/spark/examples/streaming/HdfsWordCount.scala :
val lines = ssc.textFileStream(args(0))
val words = lines.flatMap(_.split(" "))
In your case, looks like inputfile didn't correspond to an existing path.
On Fri, Jul 15, 2016 at 1:05 AM, RK Spark <[email protected]> wrote:
> val count = inputfile.flatMap(line => line.split(" ")).map(word =>
> (word,1)).reduceByKey(_ + _);
> org.apache.hadoop.mapred.InvalidInputException: Input path does not exist:
>
