I am using Spark Streaming to monitor an S3 bucket. Everything appears to
be fine. But every batch interval I get the following:

*15/04/28 16:12:36 WARN HttpMethodReleaseInputStream: Attempting to release
HttpMethod in finalize() as its response data stream has gone out of scope.
This attempt will not always succeed and cannot be relied upon! Please
ensure response data streams are always fully consumed or closed to avoid
HTTP connection starvation.*

*15/04/28 16:12:36 WARN HttpMethodReleaseInputStream: Successfully released
HttpMethod in finalize(). You were lucky this time... Please ensure
response data streams are always fully consumed or closed.*

*Traceback (most recent call last):*

*  File "/Users/vb/spark-1.3.0-bin-hadoop2.4/python/pyspark/daemon.py",
line 162, in manager*

*    code = worker(sock)*

*  File "/Users/vb/spark-1.3.0-bin-hadoop2.4/python/pyspark/daemon.py",
line 60, in worker*

*    worker_main(infile, outfile)*

*  File "/Users/vb/spark-1.3.0-bin-hadoop2.4/python/pyspark/worker.py",
line 126, in main*

*    if read_int(infile) == SpecialLengths.END_OF_STREAM:*

*  File
"/Users/vb/spark-1.3.0-bin-hadoop2.4/python/pyspark/serializers.py", line
528, in read_int*

*    raise EOFError*

*EOFError*

Does anyone know the cause of this and how to fix it?

Thanks,

Vadim
ᐧ

Reply via email to