Updating the Spark version means setting up the entire cluster once more? Or can we update it in some other way?
On Sat, Jan 17, 2015 at 3:22 PM, Akhil Das <ak...@sigmoidanalytics.com> wrote: > Can you paste the code? Also you can try updating your spark version. > > Thanks > Best Regards > > On Sat, Jan 17, 2015 at 2:40 PM, Deep Pradhan <pradhandeep1...@gmail.com> > wrote: > >> Hi, >> I am using Spark-1.0.0 in a single node cluster. When I run a job with >> small data set it runs perfectly but when I use a data set of 350 KB, no >> output is being produced and when I try to run it the second time it is >> giving me an exception telling that SparkContext was shut down. >> Can anyone help me on this? >> >> Thank You >> > >