I am using Spark Streaming to monitor an S3 bucket. Everything appears to be fine. But every batch interval I get the following:
*15/04/28 16:12:36 WARN HttpMethodReleaseInputStream: Attempting to release HttpMethod in finalize() as its response data stream has gone out of scope. This attempt will not always succeed and cannot be relied upon! Please ensure response data streams are always fully consumed or closed to avoid HTTP connection starvation.* *15/04/28 16:12:36 WARN HttpMethodReleaseInputStream: Successfully released HttpMethod in finalize(). You were lucky this time... Please ensure response data streams are always fully consumed or closed.* *Traceback (most recent call last):* * File "/Users/vb/spark-1.3.0-bin-hadoop2.4/python/pyspark/daemon.py", line 162, in manager* * code = worker(sock)* * File "/Users/vb/spark-1.3.0-bin-hadoop2.4/python/pyspark/daemon.py", line 60, in worker* * worker_main(infile, outfile)* * File "/Users/vb/spark-1.3.0-bin-hadoop2.4/python/pyspark/worker.py", line 126, in main* * if read_int(infile) == SpecialLengths.END_OF_STREAM:* * File "/Users/vb/spark-1.3.0-bin-hadoop2.4/python/pyspark/serializers.py", line 528, in read_int* * raise EOFError* *EOFError* Does anyone know the cause of this and how to fix it? Thanks, Vadim ᐧ