Deep, Yes you have another spark shell or application sticking around somewhere. Try to inspect running processes and lookout for jave process. And kill it. This might be helpful https://www.digitalocean.com/community/tutorials/how-to-use-ps-kill-and-nice-to-manage-processes-in-linux
Also, That is just a warning. FYI spark ignores BindException and probes for next available port and continues. So you application is fine if that particular error comes up. Prashant Sharma On Tue, Jan 20, 2015 at 10:30 AM, Deep Pradhan <pradhandeep1...@gmail.com> wrote: > Yes, I have increased the driver memory in spark-default.conf to 2g. Still > the error persists. > > On Tue, Jan 20, 2015 at 10:18 AM, Ted Yu <yuzhih...@gmail.com> wrote: > >> Have you seen these threads ? >> >> http://search-hadoop.com/m/JW1q5tMFlb >> http://search-hadoop.com/m/JW1q5dabji1 >> >> Cheers >> >> On Mon, Jan 19, 2015 at 8:33 PM, Deep Pradhan <pradhandeep1...@gmail.com> >> wrote: >> >>> Hi Ted, >>> When I am running the same job with small data, I am able to run. But >>> when I run it with relatively bigger set of data, it is giving me >>> OutOfMemoryError: GC overhead limit exceeded. >>> The first time I run the job, no output. When I run for second time, I >>> am getting this error. I am aware that, the memory is getting full, but is >>> there any way to avoid this? >>> I have a single node Spark cluster. >>> >>> Thank You >>> >>> On Tue, Jan 20, 2015 at 9:52 AM, Deep Pradhan <pradhandeep1...@gmail.com >>> > wrote: >>> >>>> I had the Spark Shell running through out. Is it because of that? >>>> >>>> On Tue, Jan 20, 2015 at 9:47 AM, Ted Yu <yuzhih...@gmail.com> wrote: >>>> >>>>> Was there another instance of Spark running on the same machine ? >>>>> >>>>> Can you pastebin the full stack trace ? >>>>> >>>>> Cheers >>>>> >>>>> On Mon, Jan 19, 2015 at 8:11 PM, Deep Pradhan < >>>>> pradhandeep1...@gmail.com> wrote: >>>>> >>>>>> Hi, >>>>>> I am running a Spark job. I get the output correctly but when I see >>>>>> the logs file I see the following: >>>>>> AbstractLifeCycle: FAILED.....: java.net.BindException: Address >>>>>> already in use... >>>>>> >>>>>> What could be the reason for this? >>>>>> >>>>>> Thank You >>>>>> >>>>> >>>>> >>>> >>> >> >