Re: Spark can't find jars

2014-10-27 Thread twinkle sachdeva
Hi,

Try running following in the spark folder:

bin/*run-example *SparkPi 10

If this runs fine, just see the set of arguments being passed via this
script, and try in similar way.

Thanks,


On Thu, Oct 16, 2014 at 2:59 PM, Christophe Préaud 
christophe.pre...@kelkoo.com wrote:

  Hi,

 I have created a JIRA (SPARK-3967
 https://issues.apache.org/jira/browse/SPARK-3967), can you please
 confirm that you are hit by the same issue?

 Thanks,
 Christophe.


 On 15/10/2014 09:49, Christophe Préaud wrote:

 Hi Jimmy,
 Did you try my patch?
 The problem on my side was that the hadoop.tmp.dir  (in hadoop
 core-site.xml) was not handled properly by Spark when it is set on multiple
 partitions/disks, i.e.:

 property
   namehadoop.tmp.dir/name
   value
 file:/d1/yarn/local,file:/d2/yarn/local,file:/d3/yarn/local,file:/d4/yarn/local,file:/d5/yarn/local,file:/d6/yarn/local,file:/d7/yarn/local
 /value
 /property

 Hence, you won't be hit by this bug if your hadoop.tmp.dir is set on one
 partition only.
 If your hadoop.tmp.dir is also set on several partitions, I agree that it
 looks like a bug in Spark.

 Christophe.

 On 14/10/2014 18:50, Jimmy McErlain wrote:

 So the only way that I could make this work was to build a fat jar file as
 suggested earlier.  To me (and I am no expert) it seems like this is a
 bug.  Everything was working for me prior to our upgrade to Spark 1.1 on
 Hadoop 2.2 but now it seems to not...  ie packaging my jars locally then
 pushing them out to the cluster and pointing them to corresponding
 dependent jars

  Sorry I cannot be more help!
 J
 ᐧ




  *JIMMY MCERLAIN*

 DATA SCIENTIST (NERD)

 *. . . . . . . . . . . . . . . . . .*


   *IF WE CAN’T DOUBLE YOUR SALES,*



 *ONE OF US IS IN THE WRONG BUSINESS. *

 *E*: ji...@sellpoints.com

 *M*: *510.303.7751*

 On Tue, Oct 14, 2014 at 4:59 AM, Christophe Préaud 
 christophe.pre...@kelkoo.com wrote:

  Hello,

 I have already posted a message with the exact same problem, and proposed
 a patch (the subject is Application failure in yarn-cluster mode).
 Can you test it, and see if it works for you?
 I would be glad too if someone can confirm that it is a bug in Spark
 1.1.0.

 Regards,
 Christophe.


 On 14/10/2014 03:15, Jimmy McErlain wrote:

 BTW this has always worked for me before until we upgraded the cluster to
 Spark 1.1.1...
 J
 ᐧ




  *JIMMY MCERLAIN*

 DATA SCIENTIST (NERD)

 *. . . . . . . . . . . . . . . . . .*


   *IF WE CAN’T DOUBLE YOUR SALES,*



 *ONE OF US IS IN THE WRONG BUSINESS. *

 *E*: ji...@sellpoints.com

 *M*: *510.303.7751 510.303.7751*

 On Mon, Oct 13, 2014 at 5:39 PM, HARIPRIYA AYYALASOMAYAJULA 
 aharipriy...@gmail.com wrote:

 Helo,

  Can you check if  the jar file is available in the target-scala-2.10
 folder?

  When you use sbt package to make the jar file, that is where the jar
 file would be located.

  The following command works well for me:

  spark-submit --class “Classname   --master yarn-cluster
 jarfile(withcomplete path)

 Can you try checking  with this initially and later add other options?

 On Mon, Oct 13, 2014 at 7:36 PM, Jimmy ji...@sellpoints.com wrote:

  Having the exact same error with the exact same jar Do you work
 for Altiscale? :)
 J

 Sent from my iPhone

 On Oct 13, 2014, at 5:33 PM, Andy Srine andy.sr...@gmail.com wrote:

   Hi Guys,


  Spark rookie here. I am getting a file not found exception on the
 --jars. This is on the yarn cluster mode and I am running the following
 command on our recently upgraded Spark 1.1.1 environment.


  ./bin/spark-submit --verbose --master yarn --deploy-mode cluster
 --class myEngine --driver-memory 1g --driver-library-path
 /hadoop/share/hadoop/mapreduce/lib/hadoop-lzo-0.4.18-201406111750.jar
 --executor-memory 5g --executor-cores 5 --jars
 /home/andy/spark/lib/joda-convert-1.2.jar --queue default --num-executors 4
 /home/andy/spark/lib/my-spark-lib_1.0.jar


  This is the error I am hitting. Any tips would be much appreciated.
 The file permissions looks fine on my local disk.


  14/10/13 22:49:39 INFO yarn.ApplicationMaster: Unregistering
 ApplicationMaster with FAILED

 14/10/13 22:49:39 INFO impl.AMRMClientImpl: Waiting for application to
 be successfully unregistered.

 Exception in thread Driver java.lang.reflect.InvocationTargetException

 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)

 at
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)

 at
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)

 at java.lang.reflect.Method.invoke(Method.java:606)

 at
 org.apache.spark.deploy.yarn.ApplicationMaster$$anon$2.run(ApplicationMaster.scala:162)

 Caused by: org.apache.spark.SparkException: Job aborted due to stage
 failure: Task 3 in stage 1.0 failed 4 times, most recent failure: Lost task
 3.3 in stage 1.0 (TID 12, 122-67.vb2.company.com):
 java.io.FileNotFoundException: ./joda-convert-1.2.jar (Permission denied)

 

Re: Spark can't find jars

2014-10-16 Thread Christophe Préaud
Hi,

I have created a JIRA 
(SPARK-3967https://issues.apache.org/jira/browse/SPARK-3967), can you please 
confirm that you are hit by the same issue?

Thanks,
Christophe.

On 15/10/2014 09:49, Christophe Préaud wrote:
Hi Jimmy,
Did you try my patch?
The problem on my side was that the hadoop.tmp.dir  (in hadoop core-site.xml) 
was not handled properly by Spark when it is set on multiple partitions/disks, 
i.e.:

property
  namehadoop.tmp.dir/name
  
valuefile:/d1/yarn/local,file:/d2/yarn/local,file:/d3/yarn/local,file:/d4/yarn/local,file:/d5/yarn/local,file:/d6/yarn/local,file:/d7/yarn/local/value
/property

Hence, you won't be hit by this bug if your hadoop.tmp.dir is set on one 
partition only.
If your hadoop.tmp.dir is also set on several partitions, I agree that it looks 
like a bug in Spark.

Christophe.

On 14/10/2014 18:50, Jimmy McErlain wrote:
So the only way that I could make this work was to build a fat jar file as 
suggested earlier.  To me (and I am no expert) it seems like this is a bug.  
Everything was working for me prior to our upgrade to Spark 1.1 on Hadoop 2.2 
but now it seems to not...  ie packaging my jars locally then pushing them out 
to the cluster and pointing them to corresponding dependent jars

Sorry I cannot be more help!
J
[https://mailfoogae.appspot.com/t?sender=aamltbXlAc2VsbHBvaW50cy5jb20%3Dtype=zerocontentguid=c1a21a6a-dbf9-453d-8c2a-b5e6a8d5ca56]ᐧ





JIMMY MCERLAIN

DATA SCIENTIST (NERD)

. . . . . . . . . . . . . . . . . .

[http://assetsw.sellpoint.net/IA/creative_services/logo_2014/sellpoints_logo_black_transparent_170x81.png]

IF WE CAN’T DOUBLE YOUR SALES,

ONE OF US IS IN THE WRONG BUSINESS.


E: ji...@sellpoints.commailto:ji...@sellpoints.com

M: 510.303.7751

On Tue, Oct 14, 2014 at 4:59 AM, Christophe Préaud 
christophe.pre...@kelkoo.commailto:christophe.pre...@kelkoo.com wrote:
Hello,

I have already posted a message with the exact same problem, and proposed a 
patch (the subject is Application failure in yarn-cluster mode).
Can you test it, and see if it works for you?
I would be glad too if someone can confirm that it is a bug in Spark 1.1.0.

Regards,
Christophe.


On 14/10/2014 03:15, Jimmy McErlain wrote:
BTW this has always worked for me before until we upgraded the cluster to Spark 
1.1.1...
J
[https://mailfoogae.appspot.com/t?sender=aamltbXlAc2VsbHBvaW50cy5jb20%3Dtype=zerocontentguid=92430839-642b-4921-8d42-f266e48bcdfe]ᐧ





JIMMY MCERLAIN

DATA SCIENTIST (NERD)

. . . . . . . . . . . . . . . . . .

[http://assetsw.sellpoint.net/IA/creative_services/logo_2014/sellpoints_logo_black_transparent_170x81.png]

IF WE CAN’T DOUBLE YOUR SALES,

ONE OF US IS IN THE WRONG BUSINESS.


E: ji...@sellpoints.commailto:ji...@sellpoints.com

M: 510.303.7751tel:510.303.7751

On Mon, Oct 13, 2014 at 5:39 PM, HARIPRIYA AYYALASOMAYAJULA 
aharipriy...@gmail.commailto:aharipriy...@gmail.com wrote:
Helo,

Can you check if  the jar file is available in the target-scala-2.10 folder?

When you use sbt package to make the jar file, that is where the jar file would 
be located.

The following command works well for me:


spark-submit --class “Classname   --master yarn-cluster jarfile(withcomplete 
path)

Can you try checking  with this initially and later add other options?

On Mon, Oct 13, 2014 at 7:36 PM, Jimmy 
ji...@sellpoints.commailto:ji...@sellpoints.com wrote:
Having the exact same error with the exact same jar Do you work for 
Altiscale? :)
J

Sent from my iPhone

On Oct 13, 2014, at 5:33 PM, Andy Srine 
andy.sr...@gmail.commailto:andy.sr...@gmail.com wrote:


Hi Guys,


Spark rookie here. I am getting a file not found exception on the --jars. This 
is on the yarn cluster mode and I am running the following command on our 
recently upgraded Spark 1.1.1 environment.


./bin/spark-submit --verbose --master yarn --deploy-mode cluster --class 
myEngine --driver-memory 1g --driver-library-path 
/hadoop/share/hadoop/mapreduce/lib/hadoop-lzo-0.4.18-201406111750.jar 
--executor-memory 5g --executor-cores 5 --jars 
/home/andy/spark/lib/joda-convert-1.2.jar --queue default --num-executors 4 
/home/andy/spark/lib/my-spark-lib_1.0.jar


This is the error I am hitting. Any tips would be much appreciated. The file 
permissions looks fine on my local disk.


14/10/13 22:49:39 INFO yarn.ApplicationMaster: Unregistering ApplicationMaster 
with FAILED

14/10/13 22:49:39 INFO impl.AMRMClientImpl: Waiting for application to be 
successfully unregistered.

Exception in thread Driver java.lang.reflect.InvocationTargetException

at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)

at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)

at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)

at java.lang.reflect.Method.invoke(Method.java:606)

at 
org.apache.spark.deploy.yarn.ApplicationMaster$$anon$2.run(ApplicationMaster.scala:162)

Caused by: org.apache.spark.SparkException: Job aborted due to stage failure: 

Re: Spark can't find jars

2014-10-15 Thread Christophe Préaud
Hi Jimmy,
Did you try my patch?
The problem on my side was that the hadoop.tmp.dir  (in hadoop core-site.xml) 
was not handled properly by Spark when it is set on multiple partitions/disks, 
i.e.:

property
  namehadoop.tmp.dir/name
  
valuefile:/d1/yarn/local,file:/d2/yarn/local,file:/d3/yarn/local,file:/d4/yarn/local,file:/d5/yarn/local,file:/d6/yarn/local,file:/d7/yarn/local/value
/property

Hence, you won't be hit by this bug if your hadoop.tmp.dir is set on one 
partition only.
If your hadoop.tmp.dir is also set on several partitions, I agree that it looks 
like a bug in Spark.

Christophe.

On 14/10/2014 18:50, Jimmy McErlain wrote:
So the only way that I could make this work was to build a fat jar file as 
suggested earlier.  To me (and I am no expert) it seems like this is a bug.  
Everything was working for me prior to our upgrade to Spark 1.1 on Hadoop 2.2 
but now it seems to not...  ie packaging my jars locally then pushing them out 
to the cluster and pointing them to corresponding dependent jars

Sorry I cannot be more help!
J
[https://mailfoogae.appspot.com/t?sender=aamltbXlAc2VsbHBvaW50cy5jb20%3Dtype=zerocontentguid=c1a21a6a-dbf9-453d-8c2a-b5e6a8d5ca56]ᐧ





JIMMY MCERLAIN

DATA SCIENTIST (NERD)

. . . . . . . . . . . . . . . . . .

[http://assetsw.sellpoint.net/IA/creative_services/logo_2014/sellpoints_logo_black_transparent_170x81.png]

IF WE CAN’T DOUBLE YOUR SALES,

ONE OF US IS IN THE WRONG BUSINESS.


E: ji...@sellpoints.commailto:ji...@sellpoints.com

M: 510.303.7751

On Tue, Oct 14, 2014 at 4:59 AM, Christophe Préaud 
christophe.pre...@kelkoo.commailto:christophe.pre...@kelkoo.com wrote:
Hello,

I have already posted a message with the exact same problem, and proposed a 
patch (the subject is Application failure in yarn-cluster mode).
Can you test it, and see if it works for you?
I would be glad too if someone can confirm that it is a bug in Spark 1.1.0.

Regards,
Christophe.


On 14/10/2014 03:15, Jimmy McErlain wrote:
BTW this has always worked for me before until we upgraded the cluster to Spark 
1.1.1...
J
[https://mailfoogae.appspot.com/t?sender=aamltbXlAc2VsbHBvaW50cy5jb20%3Dtype=zerocontentguid=92430839-642b-4921-8d42-f266e48bcdfe]ᐧ





JIMMY MCERLAIN

DATA SCIENTIST (NERD)

. . . . . . . . . . . . . . . . . .

[http://assetsw.sellpoint.net/IA/creative_services/logo_2014/sellpoints_logo_black_transparent_170x81.png]

IF WE CAN’T DOUBLE YOUR SALES,

ONE OF US IS IN THE WRONG BUSINESS.


E: ji...@sellpoints.commailto:ji...@sellpoints.com

M: 510.303.7751tel:510.303.7751

On Mon, Oct 13, 2014 at 5:39 PM, HARIPRIYA AYYALASOMAYAJULA 
aharipriy...@gmail.commailto:aharipriy...@gmail.com wrote:
Helo,

Can you check if  the jar file is available in the target-scala-2.10 folder?

When you use sbt package to make the jar file, that is where the jar file would 
be located.

The following command works well for me:


spark-submit --class “Classname   --master yarn-cluster jarfile(withcomplete 
path)

Can you try checking  with this initially and later add other options?

On Mon, Oct 13, 2014 at 7:36 PM, Jimmy 
ji...@sellpoints.commailto:ji...@sellpoints.com wrote:
Having the exact same error with the exact same jar Do you work for 
Altiscale? :)
J

Sent from my iPhone

On Oct 13, 2014, at 5:33 PM, Andy Srine 
andy.sr...@gmail.commailto:andy.sr...@gmail.com wrote:


Hi Guys,


Spark rookie here. I am getting a file not found exception on the --jars. This 
is on the yarn cluster mode and I am running the following command on our 
recently upgraded Spark 1.1.1 environment.


./bin/spark-submit --verbose --master yarn --deploy-mode cluster --class 
myEngine --driver-memory 1g --driver-library-path 
/hadoop/share/hadoop/mapreduce/lib/hadoop-lzo-0.4.18-201406111750.jar 
--executor-memory 5g --executor-cores 5 --jars 
/home/andy/spark/lib/joda-convert-1.2.jar --queue default --num-executors 4 
/home/andy/spark/lib/my-spark-lib_1.0.jar


This is the error I am hitting. Any tips would be much appreciated. The file 
permissions looks fine on my local disk.


14/10/13 22:49:39 INFO yarn.ApplicationMaster: Unregistering ApplicationMaster 
with FAILED

14/10/13 22:49:39 INFO impl.AMRMClientImpl: Waiting for application to be 
successfully unregistered.

Exception in thread Driver java.lang.reflect.InvocationTargetException

at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)

at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)

at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)

at java.lang.reflect.Method.invoke(Method.java:606)

at 
org.apache.spark.deploy.yarn.ApplicationMaster$$anon$2.run(ApplicationMaster.scala:162)

Caused by: org.apache.spark.SparkException: Job aborted due to stage failure: 
Task 3 in stage 1.0 failed 4 times, most recent failure: Lost task 3.3 in stage 
1.0 (TID 12, 122-67.vb2.company.comhttp://122-67.vb2.company.com): 
java.io.FileNotFoundException: ./joda-convert-1.2.jar (Permission 

Re: Spark can't find jars

2014-10-14 Thread Christophe Préaud
Hello,

I have already posted a message with the exact same problem, and proposed a 
patch (the subject is Application failure in yarn-cluster mode).
Can you test it, and see if it works for you?
I would be glad too if someone can confirm that it is a bug in Spark 1.1.0.

Regards,
Christophe.

On 14/10/2014 03:15, Jimmy McErlain wrote:
BTW this has always worked for me before until we upgraded the cluster to Spark 
1.1.1...
J
[https://mailfoogae.appspot.com/t?sender=aamltbXlAc2VsbHBvaW50cy5jb20%3Dtype=zerocontentguid=92430839-642b-4921-8d42-f266e48bcdfe]ᐧ





JIMMY MCERLAIN

DATA SCIENTIST (NERD)

. . . . . . . . . . . . . . . . . .

[http://assetsw.sellpoint.net/IA/creative_services/logo_2014/sellpoints_logo_black_transparent_170x81.png]

IF WE CAN’T DOUBLE YOUR SALES,

ONE OF US IS IN THE WRONG BUSINESS.


E: ji...@sellpoints.commailto:ji...@sellpoints.com

M: 510.303.7751

On Mon, Oct 13, 2014 at 5:39 PM, HARIPRIYA AYYALASOMAYAJULA 
aharipriy...@gmail.commailto:aharipriy...@gmail.com wrote:
Helo,

Can you check if  the jar file is available in the target-scala-2.10 folder?

When you use sbt package to make the jar file, that is where the jar file would 
be located.

The following command works well for me:


spark-submit --class “Classname   --master yarn-cluster jarfile(withcomplete 
path)

Can you try checking  with this initially and later add other options?

On Mon, Oct 13, 2014 at 7:36 PM, Jimmy 
ji...@sellpoints.commailto:ji...@sellpoints.com wrote:
Having the exact same error with the exact same jar Do you work for 
Altiscale? :)
J

Sent from my iPhone

On Oct 13, 2014, at 5:33 PM, Andy Srine 
andy.sr...@gmail.commailto:andy.sr...@gmail.com wrote:


Hi Guys,


Spark rookie here. I am getting a file not found exception on the --jars. This 
is on the yarn cluster mode and I am running the following command on our 
recently upgraded Spark 1.1.1 environment.


./bin/spark-submit --verbose --master yarn --deploy-mode cluster --class 
myEngine --driver-memory 1g --driver-library-path 
/hadoop/share/hadoop/mapreduce/lib/hadoop-lzo-0.4.18-201406111750.jar 
--executor-memory 5g --executor-cores 5 --jars 
/home/andy/spark/lib/joda-convert-1.2.jar --queue default --num-executors 4 
/home/andy/spark/lib/my-spark-lib_1.0.jar


This is the error I am hitting. Any tips would be much appreciated. The file 
permissions looks fine on my local disk.


14/10/13 22:49:39 INFO yarn.ApplicationMaster: Unregistering ApplicationMaster 
with FAILED

14/10/13 22:49:39 INFO impl.AMRMClientImpl: Waiting for application to be 
successfully unregistered.

Exception in thread Driver java.lang.reflect.InvocationTargetException

at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)

at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)

at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)

at java.lang.reflect.Method.invoke(Method.java:606)

at 
org.apache.spark.deploy.yarn.ApplicationMaster$$anon$2.run(ApplicationMaster.scala:162)

Caused by: org.apache.spark.SparkException: Job aborted due to stage failure: 
Task 3 in stage 1.0 failed 4 times, most recent failure: Lost task 3.3 in stage 
1.0 (TID 12, 122-67.vb2.company.comhttp://122-67.vb2.company.com): 
java.io.FileNotFoundException: ./joda-convert-1.2.jar (Permission denied)

java.io.FileOutputStream.open(Native Method)

java.io.FileOutputStream.init(FileOutputStream.java:221)

com.google.common.io.Files$FileByteSink.openStream(Files.java:223)

com.google.common.io.Files$FileByteSink.openStream(Files.java:211)



Thanks,
Andy




--
Regards,
Haripriya Ayyalasomayajula





Kelkoo SAS
Société par Actions Simplifiée
Au capital de € 4.168.964,30
Siège social : 8, rue du Sentier 75002 Paris
425 093 069 RCS Paris

Ce message et les pièces jointes sont confidentiels et établis à l'attention 
exclusive de leurs destinataires. Si vous n'êtes pas le destinataire de ce 
message, merci de le détruire et d'en avertir l'expéditeur.


Re: Spark can't find jars

2014-10-14 Thread Jimmy McErlain
So the only way that I could make this work was to build a fat jar file as
suggested earlier.  To me (and I am no expert) it seems like this is a
bug.  Everything was working for me prior to our upgrade to Spark 1.1 on
Hadoop 2.2 but now it seems to not...  ie packaging my jars locally then
pushing them out to the cluster and pointing them to corresponding
dependent jars

Sorry I cannot be more help!
J
ᐧ




*JIMMY MCERLAIN*

DATA SCIENTIST (NERD)

*. . . . . . . . . . . . . . . . . .*


*IF WE CAN’T DOUBLE YOUR SALES,*



*ONE OF US IS IN THE WRONG BUSINESS.*

*E*: ji...@sellpoints.com

*M*: *510.303.7751*

On Tue, Oct 14, 2014 at 4:59 AM, Christophe Préaud 
christophe.pre...@kelkoo.com wrote:

  Hello,

 I have already posted a message with the exact same problem, and proposed
 a patch (the subject is Application failure in yarn-cluster mode).
 Can you test it, and see if it works for you?
 I would be glad too if someone can confirm that it is a bug in Spark 1.1.0.

 Regards,
 Christophe.


 On 14/10/2014 03:15, Jimmy McErlain wrote:

 BTW this has always worked for me before until we upgraded the cluster to
 Spark 1.1.1...
 J
 ᐧ




  *JIMMY MCERLAIN*

 DATA SCIENTIST (NERD)

 *. . . . . . . . . . . . . . . . . .*


   *IF WE CAN’T DOUBLE YOUR SALES,*



 *ONE OF US IS IN THE WRONG BUSINESS. *

 *E*: ji...@sellpoints.com

 *M*: *510.303.7751 510.303.7751*

 On Mon, Oct 13, 2014 at 5:39 PM, HARIPRIYA AYYALASOMAYAJULA 
 aharipriy...@gmail.com wrote:

 Helo,

  Can you check if  the jar file is available in the target-scala-2.10
 folder?

  When you use sbt package to make the jar file, that is where the jar
 file would be located.

  The following command works well for me:

  spark-submit --class “Classname   --master yarn-cluster
 jarfile(withcomplete path)

 Can you try checking  with this initially and later add other options?

 On Mon, Oct 13, 2014 at 7:36 PM, Jimmy ji...@sellpoints.com wrote:

  Having the exact same error with the exact same jar Do you work
 for Altiscale? :)
 J

 Sent from my iPhone

 On Oct 13, 2014, at 5:33 PM, Andy Srine andy.sr...@gmail.com wrote:

   Hi Guys,


  Spark rookie here. I am getting a file not found exception on the
 --jars. This is on the yarn cluster mode and I am running the following
 command on our recently upgraded Spark 1.1.1 environment.


  ./bin/spark-submit --verbose --master yarn --deploy-mode cluster
 --class myEngine --driver-memory 1g --driver-library-path
 /hadoop/share/hadoop/mapreduce/lib/hadoop-lzo-0.4.18-201406111750.jar
 --executor-memory 5g --executor-cores 5 --jars
 /home/andy/spark/lib/joda-convert-1.2.jar --queue default --num-executors 4
 /home/andy/spark/lib/my-spark-lib_1.0.jar


  This is the error I am hitting. Any tips would be much appreciated.
 The file permissions looks fine on my local disk.


  14/10/13 22:49:39 INFO yarn.ApplicationMaster: Unregistering
 ApplicationMaster with FAILED

 14/10/13 22:49:39 INFO impl.AMRMClientImpl: Waiting for application to
 be successfully unregistered.

 Exception in thread Driver java.lang.reflect.InvocationTargetException

 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)

 at
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)

 at
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)

 at java.lang.reflect.Method.invoke(Method.java:606)

 at
 org.apache.spark.deploy.yarn.ApplicationMaster$$anon$2.run(ApplicationMaster.scala:162)

 Caused by: org.apache.spark.SparkException: Job aborted due to stage
 failure: Task 3 in stage 1.0 failed 4 times, most recent failure: Lost task
 3.3 in stage 1.0 (TID 12, 122-67.vb2.company.com):
 java.io.FileNotFoundException: ./joda-convert-1.2.jar (Permission denied)

 java.io.FileOutputStream.open(Native Method)

 java.io.FileOutputStream.init(FileOutputStream.java:221)


 com.google.common.io.Files$FileByteSink.openStream(Files.java:223)


 com.google.common.io.Files$FileByteSink.openStream(Files.java:211)


 Thanks,
 Andy




   --
 Regards,
 Haripriya Ayyalasomayajula




 --
 Kelkoo SAS
 Société par Actions Simplifiée
 Au capital de € 4.168.964,30
 Siège social : 8, rue du Sentier 75002 Paris
 425 093 069 RCS Paris

 Ce message et les pièces jointes sont confidentiels et établis à
 l'attention exclusive de leurs destinataires. Si vous n'êtes pas le
 destinataire de ce message, merci de le détruire et d'en avertir
 l'expéditeur.



Spark can't find jars

2014-10-13 Thread Andy Srine
Hi Guys,


Spark rookie here. I am getting a file not found exception on the --jars.
This is on the yarn cluster mode and I am running the following command on
our recently upgraded Spark 1.1.1 environment.


./bin/spark-submit --verbose --master yarn --deploy-mode cluster --class
myEngine --driver-memory 1g --driver-library-path
/hadoop/share/hadoop/mapreduce/lib/hadoop-lzo-0.4.18-201406111750.jar
--executor-memory 5g --executor-cores 5 --jars
/home/andy/spark/lib/joda-convert-1.2.jar --queue default --num-executors 4
/home/andy/spark/lib/my-spark-lib_1.0.jar


This is the error I am hitting. Any tips would be much appreciated. The
file permissions looks fine on my local disk.


14/10/13 22:49:39 INFO yarn.ApplicationMaster: Unregistering
ApplicationMaster with FAILED

14/10/13 22:49:39 INFO impl.AMRMClientImpl: Waiting for application to be
successfully unregistered.

Exception in thread Driver java.lang.reflect.InvocationTargetException

at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)

at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)

at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)

at java.lang.reflect.Method.invoke(Method.java:606)

at
org.apache.spark.deploy.yarn.ApplicationMaster$$anon$2.run(ApplicationMaster.scala:162)

Caused by: org.apache.spark.SparkException: Job aborted due to stage
failure: Task 3 in stage 1.0 failed 4 times, most recent failure: Lost task
3.3 in stage 1.0 (TID 12, 122-67.vb2.company.com):
java.io.FileNotFoundException: ./joda-convert-1.2.jar (Permission denied)

java.io.FileOutputStream.open(Native Method)

java.io.FileOutputStream.init(FileOutputStream.java:221)

com.google.common.io.Files$FileByteSink.openStream(Files.java:223)

com.google.common.io.Files$FileByteSink.openStream(Files.java:211)


Thanks,
Andy


Re: Spark can't find jars

2014-10-13 Thread Jimmy
Having the exact same error with the exact same jar Do you work for 
Altiscale? :) 
J

Sent from my iPhone

 On Oct 13, 2014, at 5:33 PM, Andy Srine andy.sr...@gmail.com wrote:
 
 Hi Guys,
 
 Spark rookie here. I am getting a file not found exception on the --jars. 
 This is on the yarn cluster mode and I am running the following command on 
 our recently upgraded Spark 1.1.1 environment.
 
 ./bin/spark-submit --verbose --master yarn --deploy-mode cluster --class 
 myEngine --driver-memory 1g --driver-library-path 
 /hadoop/share/hadoop/mapreduce/lib/hadoop-lzo-0.4.18-201406111750.jar 
 --executor-memory 5g --executor-cores 5 --jars 
 /home/andy/spark/lib/joda-convert-1.2.jar --queue default --num-executors 4 
 /home/andy/spark/lib/my-spark-lib_1.0.jar
 
 This is the error I am hitting. Any tips would be much appreciated. The file 
 permissions looks fine on my local disk.
 
 14/10/13 22:49:39 INFO yarn.ApplicationMaster: Unregistering 
 ApplicationMaster with FAILED
 14/10/13 22:49:39 INFO impl.AMRMClientImpl: Waiting for application to be 
 successfully unregistered.
 Exception in thread Driver java.lang.reflect.InvocationTargetException
   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
   at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
   at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
   at java.lang.reflect.Method.invoke(Method.java:606)
   at 
 org.apache.spark.deploy.yarn.ApplicationMaster$$anon$2.run(ApplicationMaster.scala:162)
 Caused by: org.apache.spark.SparkException: Job aborted due to stage failure: 
 Task 3 in stage 1.0 failed 4 times, most recent failure: Lost task 3.3 in 
 stage 1.0 (TID 12, 122-67.vb2.company.com): java.io.FileNotFoundException: 
 ./joda-convert-1.2.jar (Permission denied)
 java.io.FileOutputStream.open(Native Method)
 java.io.FileOutputStream.init(FileOutputStream.java:221)
 com.google.common.io.Files$FileByteSink.openStream(Files.java:223)
 com.google.common.io.Files$FileByteSink.openStream(Files.java:211)
 
 Thanks,
 Andy
 


Re: Spark can't find jars

2014-10-13 Thread Jimmy McErlain
BTW this has always worked for me before until we upgraded the cluster to
Spark 1.1.1...
J
ᐧ




*JIMMY MCERLAIN*

DATA SCIENTIST (NERD)

*. . . . . . . . . . . . . . . . . .*


*IF WE CAN’T DOUBLE YOUR SALES,*



*ONE OF US IS IN THE WRONG BUSINESS.*

*E*: ji...@sellpoints.com

*M*: *510.303.7751*

On Mon, Oct 13, 2014 at 5:39 PM, HARIPRIYA AYYALASOMAYAJULA 
aharipriy...@gmail.com wrote:

 Helo,

 Can you check if  the jar file is available in the target-scala-2.10
 folder?

 When you use sbt package to make the jar file, that is where the jar file
 would be located.

 The following command works well for me:

 spark-submit --class “Classname   --master yarn-cluster
 jarfile(withcomplete path)

 Can you try checking  with this initially and later add other options?

 On Mon, Oct 13, 2014 at 7:36 PM, Jimmy ji...@sellpoints.com wrote:

 Having the exact same error with the exact same jar Do you work for
 Altiscale? :)
 J

 Sent from my iPhone

 On Oct 13, 2014, at 5:33 PM, Andy Srine andy.sr...@gmail.com wrote:

 Hi Guys,


 Spark rookie here. I am getting a file not found exception on the --jars.
 This is on the yarn cluster mode and I am running the following command on
 our recently upgraded Spark 1.1.1 environment.


 ./bin/spark-submit --verbose --master yarn --deploy-mode cluster --class
 myEngine --driver-memory 1g --driver-library-path
 /hadoop/share/hadoop/mapreduce/lib/hadoop-lzo-0.4.18-201406111750.jar
 --executor-memory 5g --executor-cores 5 --jars
 /home/andy/spark/lib/joda-convert-1.2.jar --queue default --num-executors 4
 /home/andy/spark/lib/my-spark-lib_1.0.jar


 This is the error I am hitting. Any tips would be much appreciated. The
 file permissions looks fine on my local disk.


 14/10/13 22:49:39 INFO yarn.ApplicationMaster: Unregistering
 ApplicationMaster with FAILED

 14/10/13 22:49:39 INFO impl.AMRMClientImpl: Waiting for application to be
 successfully unregistered.

 Exception in thread Driver java.lang.reflect.InvocationTargetException

 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)

 at
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)

 at
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)

 at java.lang.reflect.Method.invoke(Method.java:606)

 at
 org.apache.spark.deploy.yarn.ApplicationMaster$$anon$2.run(ApplicationMaster.scala:162)

 Caused by: org.apache.spark.SparkException: Job aborted due to stage
 failure: Task 3 in stage 1.0 failed 4 times, most recent failure: Lost task
 3.3 in stage 1.0 (TID 12, 122-67.vb2.company.com):
 java.io.FileNotFoundException: ./joda-convert-1.2.jar (Permission denied)

 java.io.FileOutputStream.open(Native Method)

 java.io.FileOutputStream.init(FileOutputStream.java:221)

 com.google.common.io.Files$FileByteSink.openStream(Files.java:223)

 com.google.common.io.Files$FileByteSink.openStream(Files.java:211)


 Thanks,
 Andy




 --
 Regards,
 Haripriya Ayyalasomayajula




Re: Spark can't find jars

2014-10-13 Thread HARIPRIYA AYYALASOMAYAJULA
Well in the cluster, can you try copying the entire folder and then run?
For example my home folder say helloWorld consists of the src, target etc.
can you copy the entire folder in the cluster ?  I doubt it is looking for
some dependencies and is missing that when it runs your jar file.
or if you mention some input file which your program uses, ensure that it
is in hdfs and correct path is specified.

These were a few things I checked when I got a filenotfound exception

On Monday, October 13, 2014, Jimmy McErlain ji...@sellpoints.com wrote:

 That didnt seem to work... the jar files are in the target  scala2.10
 folder when I package, then I move the jar to the cluster and launch the
 app... still the same error...  Thoughts?
 J
 ᐧ




 *JIMMY MCERLAIN*

 DATA SCIENTIST (NERD)

 *. . . . . . . . . . . . . . . . . .*


 *IF WE CAN’T DOUBLE YOUR SALES,*



 *ONE OF US IS IN THE WRONG BUSINESS.*

 *E*: ji...@sellpoints.com
 javascript:_e(%7B%7D,'cvml','ji...@sellpoints.com');

 *M*: *510.303.7751*

 On Mon, Oct 13, 2014 at 5:39 PM, HARIPRIYA AYYALASOMAYAJULA 
 aharipriy...@gmail.com
 javascript:_e(%7B%7D,'cvml','aharipriy...@gmail.com'); wrote:

 Helo,

 Can you check if  the jar file is available in the target-scala-2.10
 folder?

 When you use sbt package to make the jar file, that is where the jar file
 would be located.

 The following command works well for me:

 spark-submit --class “Classname   --master yarn-cluster
 jarfile(withcomplete path)

 Can you try checking  with this initially and later add other options?

 On Mon, Oct 13, 2014 at 7:36 PM, Jimmy ji...@sellpoints.com
 javascript:_e(%7B%7D,'cvml','ji...@sellpoints.com'); wrote:

 Having the exact same error with the exact same jar Do you work for
 Altiscale? :)
 J

 Sent from my iPhone

 On Oct 13, 2014, at 5:33 PM, Andy Srine andy.sr...@gmail.com
 javascript:_e(%7B%7D,'cvml','andy.sr...@gmail.com'); wrote:

 Hi Guys,


 Spark rookie here. I am getting a file not found exception on the
 --jars. This is on the yarn cluster mode and I am running the following
 command on our recently upgraded Spark 1.1.1 environment.


 ./bin/spark-submit --verbose --master yarn --deploy-mode cluster --class
 myEngine --driver-memory 1g --driver-library-path
 /hadoop/share/hadoop/mapreduce/lib/hadoop-lzo-0.4.18-201406111750.jar
 --executor-memory 5g --executor-cores 5 --jars
 /home/andy/spark/lib/joda-convert-1.2.jar --queue default --num-executors 4
 /home/andy/spark/lib/my-spark-lib_1.0.jar


 This is the error I am hitting. Any tips would be much appreciated. The
 file permissions looks fine on my local disk.


 14/10/13 22:49:39 INFO yarn.ApplicationMaster: Unregistering
 ApplicationMaster with FAILED

 14/10/13 22:49:39 INFO impl.AMRMClientImpl: Waiting for application to
 be successfully unregistered.

 Exception in thread Driver java.lang.reflect.InvocationTargetException

 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)

 at
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)

 at
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)

 at java.lang.reflect.Method.invoke(Method.java:606)

 at
 org.apache.spark.deploy.yarn.ApplicationMaster$$anon$2.run(ApplicationMaster.scala:162)

 Caused by: org.apache.spark.SparkException: Job aborted due to stage
 failure: Task 3 in stage 1.0 failed 4 times, most recent failure: Lost task
 3.3 in stage 1.0 (TID 12, 122-67.vb2.company.com):
 java.io.FileNotFoundException: ./joda-convert-1.2.jar (Permission denied)

 java.io.FileOutputStream.open(Native Method)

 java.io.FileOutputStream.init(FileOutputStream.java:221)


 com.google.common.io.Files$FileByteSink.openStream(Files.java:223)


 com.google.common.io.Files$FileByteSink.openStream(Files.java:211)


 Thanks,
 Andy




 --
 Regards,
 Haripriya Ayyalasomayajula




-- 
Regards,
Haripriya Ayyalasomayajula
contact : 650-796-7112


Re: Spark can't find jars

2014-10-13 Thread HARIPRIYA AYYALASOMAYAJULA
Or if it has something to do with the way you package your files - try
another alternative method and see if it works

On Monday, October 13, 2014, HARIPRIYA AYYALASOMAYAJULA 
aharipriy...@gmail.com wrote:

 Well in the cluster, can you try copying the entire folder and then run?
 For example my home folder say helloWorld consists of the src, target etc.
 can you copy the entire folder in the cluster ?  I doubt it is looking for
 some dependencies and is missing that when it runs your jar file.
 or if you mention some input file which your program uses, ensure that it
 is in hdfs and correct path is specified.

 These were a few things I checked when I got a filenotfound exception

 On Monday, October 13, 2014, Jimmy McErlain ji...@sellpoints.com
 javascript:_e(%7B%7D,'cvml','ji...@sellpoints.com'); wrote:

 That didnt seem to work... the jar files are in the target  scala2.10
 folder when I package, then I move the jar to the cluster and launch the
 app... still the same error...  Thoughts?
 J
 ᐧ




 *JIMMY MCERLAIN*

 DATA SCIENTIST (NERD)

 *. . . . . . . . . . . . . . . . . .*


 *IF WE CAN’T DOUBLE YOUR SALES,*



 *ONE OF US IS IN THE WRONG BUSINESS.*

 *E*: ji...@sellpoints.com

 *M*: *510.303.7751*

 On Mon, Oct 13, 2014 at 5:39 PM, HARIPRIYA AYYALASOMAYAJULA 
 aharipriy...@gmail.com wrote:

 Helo,

 Can you check if  the jar file is available in the target-scala-2.10
 folder?

 When you use sbt package to make the jar file, that is where the jar
 file would be located.

 The following command works well for me:

 spark-submit --class “Classname   --master yarn-cluster
 jarfile(withcomplete path)

 Can you try checking  with this initially and later add other options?

 On Mon, Oct 13, 2014 at 7:36 PM, Jimmy ji...@sellpoints.com wrote:

 Having the exact same error with the exact same jar Do you work for
 Altiscale? :)
 J

 Sent from my iPhone

 On Oct 13, 2014, at 5:33 PM, Andy Srine andy.sr...@gmail.com wrote:

 Hi Guys,


 Spark rookie here. I am getting a file not found exception on the
 --jars. This is on the yarn cluster mode and I am running the following
 command on our recently upgraded Spark 1.1.1 environment.


 ./bin/spark-submit --verbose --master yarn --deploy-mode cluster
 --class myEngine --driver-memory 1g --driver-library-path
 /hadoop/share/hadoop/mapreduce/lib/hadoop-lzo-0.4.18-201406111750.jar
 --executor-memory 5g --executor-cores 5 --jars
 /home/andy/spark/lib/joda-convert-1.2.jar --queue default --num-executors 4
 /home/andy/spark/lib/my-spark-lib_1.0.jar


 This is the error I am hitting. Any tips would be much appreciated. The
 file permissions looks fine on my local disk.


 14/10/13 22:49:39 INFO yarn.ApplicationMaster: Unregistering
 ApplicationMaster with FAILED

 14/10/13 22:49:39 INFO impl.AMRMClientImpl: Waiting for application to
 be successfully unregistered.

 Exception in thread Driver java.lang.reflect.InvocationTargetException

 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)

 at
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)

 at
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)

 at java.lang.reflect.Method.invoke(Method.java:606)

 at
 org.apache.spark.deploy.yarn.ApplicationMaster$$anon$2.run(ApplicationMaster.scala:162)

 Caused by: org.apache.spark.SparkException: Job aborted due to stage
 failure: Task 3 in stage 1.0 failed 4 times, most recent failure: Lost task
 3.3 in stage 1.0 (TID 12, 122-67.vb2.company.com):
 java.io.FileNotFoundException: ./joda-convert-1.2.jar (Permission denied)

 java.io.FileOutputStream.open(Native Method)

 java.io.FileOutputStream.init(FileOutputStream.java:221)


 com.google.common.io.Files$FileByteSink.openStream(Files.java:223)


 com.google.common.io.Files$FileByteSink.openStream(Files.java:211)


 Thanks,
 Andy




 --
 Regards,
 Haripriya Ayyalasomayajula




 --
 Regards,
 Haripriya Ayyalasomayajula
 contact : 650-796-7112



-- 
Regards,
Haripriya Ayyalasomayajula
contact : 650-796-7112