Hi, I experienced exactly the same problems when using SparkContext with "local[1]" master specification, because in that case one thread is used for receiving data, the others for processing. As there is only one thread running, no processing will take place. Once you shut down the connection, the receiver thread will be used for processing.
Any chance you run into the same issue? Tobias On Mon, Jul 14, 2014 at 11:45 AM, kytay <kaiyang....@gmail.com> wrote: > Hi Akhil Das > > Thanks. > > I tried the codes. and it works. > > There's a problem with my socket codes that is not flushing the content > out, > and for the test tool, Hercules, I have to close the socket connection to > "flush" the content out. > > I am going to troubleshoot why nc works, and the codes and test tool don't. > > > > > > -- > View this message in context: > http://apache-spark-user-list.1001560.n3.nabble.com/Streaming-Cannot-get-socketTextStream-to-receive-anything-tp9382p9576.html > Sent from the Apache Spark User List mailing list archive at Nabble.com. >