Hi, Thanks for your answer.
I have chose "Spark" in the "job type". There is not any option where we can choose the version. How I can choose different version? Thanks, Anahita On Thu, Jan 12, 2017 at 6:39 PM, A Shaikh <shaikh.af...@gmail.com> wrote: > You may have tested this code on Spark version on your local machine > version of which may be different to whats in Google Cloud Storage. > You need to select appropraite Spark version when you submit your job. > > On 12 January 2017 at 15:51, Anahita Talebi <anahita.t.am...@gmail.com> > wrote: > >> Dear all, >> >> I am trying to run a .jar file as a job using submit job in google cloud >> console. >> https://cloud.google.com/dataproc/docs/guides/submit-job >> >> I actually ran the spark code on my local computer to generate a .jar >> file. Then in the Argument folder, I give the value of the arguments that I >> used in the spark code. One of the argument is training data set that I put >> in the same bucket that I save my .jar file. In the bucket, I put only the >> .jar file, training dataset and testing dataset. >> >> Main class or jar >> gs://Anahita/test.jar >> >> Arguments >> >> --lambda=.001 >> --eta=1.0 >> --trainFile=gs://Anahita/small_train.dat >> --testFile=gs://Anahita/small_test.dat >> >> The problem is that when I run the job I get the following error and >> actually it cannot read my training and testing data sets. >> >> Exception in thread "main" java.lang.NoSuchMethodError: >> org.apache.spark.rdd.RDD.coalesce(IZLscala/math/Ordering;)Lorg/apache/spark/rdd/RDD; >> >> Can anyone help me how I can solve this problem? >> >> Thanks, >> >> Anahita >> >> >> >