RE: no space left at worker node

2015-02-09 Thread ey-chih chow
I change to submit command to the following, but the jar file is still copied 
to the directory of ./spark/work/app-xx-xx.
/root/spark/bin/spark-submit --class com.crowdstar.etl.ParseAndClean --master 
spark://ec2-54-213-73-150.us-west-2.compute.amazonaws.com:7077 
local:///root/etl-admin/jar/spark-etl-0.0.1-SNAPSHOT.jar 
s3://pixlogstxt/ETL/input/2015/01/29/00/ 
file:///root/etl-admin/vertica/VERTICA.avdl 
file:///root/etl-admin/vertica/extras.json 
file:///root/etl-admin/jar/spark-etl-0.0.1-SNAPSHOT.jar 
s3://pixlogstxt/ETL/output/ -1

The content of the directory /root/spark/work/app-20150209210707-0007/0 is:
spark-etl-0.0.1-SNAPSHOT.jar  stderr  stdout

Is there any way we can disallow the system to copy the jar file?

Ey-Chih Chow 

To: 2dot7kelvin@gmail.comCC: gen.tan...@gmail.com; user@spark.apache.org
Subject: RE: no space left at worker node
Date: Mon, 9 Feb 2015 12:07:17 -0800




In other words, the working command is:
/root/spark/bin/spark-submit --class com.crowdstar.etl.ParseAndClean --master 
spark://ec2-54-213-73-150.us-west-2.compute.amazonaws.com:7077 --deploy-mode 
cluster --total-executor-cores 4 
file:///root/etl-admin/jar/spark-etl-0.0.1-SNAPSHOT.jar 
s3://pixlogstxt/ETL/input/2015/01/28/09/ 
file:///root/etl-admin/vertica/VERTICA.avdl 
file:///root/etl-admin/vertica/extras.json 
file:///root/etl-admin/jar/spark-etl-0.0.1-SNAPSHOT.jar 
s3://pixlogstxt/ETL/output/ -1
How can I change it to avoid copying the jar file to ./spark/work/app- ?
Ey-Chih ChowFrom: eyc...@hotmail.com
To: 2dot7kel...@gmail.com
CC: gen.tan...@gmail.com; user@spark.apache.org
Subject: RE: no space left at worker node
Date: Mon, 9 Feb 2015 10:59:00 -0800




Thanks.  But, in spark-submit, I specified the jar file in the form of 
local:/spark-etl-0.0.1-SNAPSHOT.jar.  It comes back with the following.  What's 
wrong with this?  
Ey-Chih Chow

===

Date: Sun, 8 Feb 2015 22:27:17 -0800Sending launch command to 
spark://ec2-54-213-73-150.us-west-2.compute.amazonaws.com:7077Driver 
successfully submitted as driver-20150209185453-0010... waiting before polling 
master for driver state... polling master for driver stateState of 
driver-20150209185453-0010 is ERRORException from cluster was: 
java.io.IOException: No FileSystem for scheme: localjava.io.IOException: No 
FileSystem for scheme: localat 
org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:1383)   at 
org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:66)   at 
org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:1404)  at 
org.apache.hadoop.fs.FileSystem.get(FileSystem.java:254) at 
org.apache.hadoop.fs.Path.getFileSystem(Path.java:187)   at 
org.apache.spark.deploy.worker.DriverRunner.org$apache$spark$deploy$worker$DriverRunner$$downloadUserJar(DriverRunner.scala:148)
 at 
org.apache.spark.deploy.worker.DriverRunner$$anon$1.run(DriverRunner.scala:74)



Subject: Re: no space left at worker node
From: 2dot7kel...@gmail.com
To: eyc...@hotmail.com
CC: gen.tan...@gmail.com; user@spark.apache.org

Maybe, try with "local:" under the heading of Advanced Dependency Management 
here: https://spark.apache.org/docs/1.1.0/submitting-applications.html
It seems this is what you want. Hope this help.
Kelvin
On Sun, Feb 8, 2015 at 9:13 PM, ey-chih chow  wrote:



Is there any way we can disable Spark copying the jar file to the corresponding 
directory.  I have a fat jar and is already copied to worker nodes using the 
command copydir.  Why Spark needs to save the jar to ./spark/work/appid each 
time a job get started?
Ey-Chih Chow 

Date: Sun, 8 Feb 2015 20:09:32 -0800
Subject: Re: no space left at worker node
From: 2dot7kel...@gmail.com
To: eyc...@hotmail.com
CC: gen.tan...@gmail.com; user@spark.apache.org

I guess you may set the parameters below to clean the directories:
spark.worker.cleanup.enabledspark.worker.cleanup.intervalspark.worker.cleanup.appDataTtl
They are described here: 
http://spark.apache.org/docs/1.2.0/spark-standalone.html
Kelvin
On Sun, Feb 8, 2015 at 5:15 PM, ey-chih chow  wrote:



I found the problem is, for each application, the Spark worker node saves the 
corresponding std output and std err under ./spark/work/appid, where appid is 
the id of the application.  If I ran several applications in a row, it will out 
of space.  In my case, the disk usage under ./spark/work/ is as follows:
1689784 ./app-20150208203033-0002/01689788  ./app-20150208203033-000240324  
./driver-20150208180505-00011691400 ./app-20150208180509-0001/01691404  
./app-20150208180509-000140316  ./driver-20150208203030-000240320   
./driver-20150208173156-1649876 ./app-20150208173200-/01649880  
./app-20150208173200-5152036.
Any suggestion how to resolve it?  Thanks.
Ey-Chih ChowFrom: eyc...@hotmail.com
To: gen.tan...@gmail.com
CC: user@spark.apache.org
Subject: RE: no space left at worker node
Date: Sun, 8 Feb 2015 15:25:43 -0800




B

RE: no space left at worker node

2015-02-09 Thread ey-chih chow
In other words, the working command is:
/root/spark/bin/spark-submit --class com.crowdstar.etl.ParseAndClean --master 
spark://ec2-54-213-73-150.us-west-2.compute.amazonaws.com:7077 --deploy-mode 
cluster --total-executor-cores 4 
file:///root/etl-admin/jar/spark-etl-0.0.1-SNAPSHOT.jar 
s3://pixlogstxt/ETL/input/2015/01/28/09/ 
file:///root/etl-admin/vertica/VERTICA.avdl 
file:///root/etl-admin/vertica/extras.json 
file:///root/etl-admin/jar/spark-etl-0.0.1-SNAPSHOT.jar 
s3://pixlogstxt/ETL/output/ -1
How can I change it to avoid copying the jar file to ./spark/work/app- ?
Ey-Chih ChowFrom: eyc...@hotmail.com
To: 2dot7kel...@gmail.com
CC: gen.tan...@gmail.com; user@spark.apache.org
Subject: RE: no space left at worker node
Date: Mon, 9 Feb 2015 10:59:00 -0800




Thanks.  But, in spark-submit, I specified the jar file in the form of 
local:/spark-etl-0.0.1-SNAPSHOT.jar.  It comes back with the following.  What's 
wrong with this?  
Ey-Chih Chow

===

Date: Sun, 8 Feb 2015 22:27:17 -0800Sending launch command to 
spark://ec2-54-213-73-150.us-west-2.compute.amazonaws.com:7077Driver 
successfully submitted as driver-20150209185453-0010... waiting before polling 
master for driver state... polling master for driver stateState of 
driver-20150209185453-0010 is ERRORException from cluster was: 
java.io.IOException: No FileSystem for scheme: localjava.io.IOException: No 
FileSystem for scheme: localat 
org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:1383)   at 
org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:66)   at 
org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:1404)  at 
org.apache.hadoop.fs.FileSystem.get(FileSystem.java:254) at 
org.apache.hadoop.fs.Path.getFileSystem(Path.java:187)   at 
org.apache.spark.deploy.worker.DriverRunner.org$apache$spark$deploy$worker$DriverRunner$$downloadUserJar(DriverRunner.scala:148)
 at 
org.apache.spark.deploy.worker.DriverRunner$$anon$1.run(DriverRunner.scala:74)



Subject: Re: no space left at worker node
From: 2dot7kel...@gmail.com
To: eyc...@hotmail.com
CC: gen.tan...@gmail.com; user@spark.apache.org

Maybe, try with "local:" under the heading of Advanced Dependency Management 
here: https://spark.apache.org/docs/1.1.0/submitting-applications.html
It seems this is what you want. Hope this help.
Kelvin
On Sun, Feb 8, 2015 at 9:13 PM, ey-chih chow  wrote:



Is there any way we can disable Spark copying the jar file to the corresponding 
directory.  I have a fat jar and is already copied to worker nodes using the 
command copydir.  Why Spark needs to save the jar to ./spark/work/appid each 
time a job get started?
Ey-Chih Chow 

Date: Sun, 8 Feb 2015 20:09:32 -0800
Subject: Re: no space left at worker node
From: 2dot7kel...@gmail.com
To: eyc...@hotmail.com
CC: gen.tan...@gmail.com; user@spark.apache.org

I guess you may set the parameters below to clean the directories:
spark.worker.cleanup.enabledspark.worker.cleanup.intervalspark.worker.cleanup.appDataTtl
They are described here: 
http://spark.apache.org/docs/1.2.0/spark-standalone.html
Kelvin
On Sun, Feb 8, 2015 at 5:15 PM, ey-chih chow  wrote:



I found the problem is, for each application, the Spark worker node saves the 
corresponding std output and std err under ./spark/work/appid, where appid is 
the id of the application.  If I ran several applications in a row, it will out 
of space.  In my case, the disk usage under ./spark/work/ is as follows:
1689784 ./app-20150208203033-0002/01689788  ./app-20150208203033-000240324  
./driver-20150208180505-00011691400 ./app-20150208180509-0001/01691404  
./app-20150208180509-000140316  ./driver-20150208203030-000240320   
./driver-20150208173156-1649876 ./app-20150208173200-/01649880  
./app-20150208173200-5152036.
Any suggestion how to resolve it?  Thanks.
Ey-Chih ChowFrom: eyc...@hotmail.com
To: gen.tan...@gmail.com
CC: user@spark.apache.org
Subject: RE: no space left at worker node
Date: Sun, 8 Feb 2015 15:25:43 -0800




By this way, the input and output paths of the job are all in s3.  I did not 
use paths of hdfs as input or output.
Best regards,
Ey-Chih Chow

From: eyc...@hotmail.com
To: gen.tan...@gmail.com
CC: user@spark.apache.org
Subject: RE: no space left at worker node
Date: Sun, 8 Feb 2015 14:57:15 -0800




Hi Gen,
Thanks.  I save my logs in a file under /var/log.  This is the only place to 
save data.  Will the problem go away if I use a better machine?
Best regards,
Ey-Chih Chow

Date: Sun, 8 Feb 2015 23:32:27 +0100
Subject: Re: no space left at worker node
From: gen.tan...@gmail.com
To: eyc...@hotmail.com
CC: user@spark.apache.org

Hi,
I am sorry that I made a mistake. r3.large has only one SSD which has been 
mounted in /mnt. Therefore this is no /dev/sdc.In fact, the problem is that 
there is no space in the under / directory. So you should check whether your 
application write data under thi

RE: no space left at worker node

2015-02-09 Thread ey-chih chow
Thanks.  But, in spark-submit, I specified the jar file in the form of 
local:/spark-etl-0.0.1-SNAPSHOT.jar.  It comes back with the following.  What's 
wrong with this?  
Ey-Chih Chow

===

Date: Sun, 8 Feb 2015 22:27:17 -0800Sending launch command to 
spark://ec2-54-213-73-150.us-west-2.compute.amazonaws.com:7077Driver 
successfully submitted as driver-20150209185453-0010... waiting before polling 
master for driver state... polling master for driver stateState of 
driver-20150209185453-0010 is ERRORException from cluster was: 
java.io.IOException: No FileSystem for scheme: localjava.io.IOException: No 
FileSystem for scheme: localat 
org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:1383)   at 
org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:66)   at 
org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:1404)  at 
org.apache.hadoop.fs.FileSystem.get(FileSystem.java:254) at 
org.apache.hadoop.fs.Path.getFileSystem(Path.java:187)   at 
org.apache.spark.deploy.worker.DriverRunner.org$apache$spark$deploy$worker$DriverRunner$$downloadUserJar(DriverRunner.scala:148)
 at 
org.apache.spark.deploy.worker.DriverRunner$$anon$1.run(DriverRunner.scala:74)



Subject: Re: no space left at worker node
From: 2dot7kel...@gmail.com
To: eyc...@hotmail.com
CC: gen.tan...@gmail.com; user@spark.apache.org

Maybe, try with "local:" under the heading of Advanced Dependency Management 
here: https://spark.apache.org/docs/1.1.0/submitting-applications.html
It seems this is what you want. Hope this help.
Kelvin
On Sun, Feb 8, 2015 at 9:13 PM, ey-chih chow  wrote:



Is there any way we can disable Spark copying the jar file to the corresponding 
directory.  I have a fat jar and is already copied to worker nodes using the 
command copydir.  Why Spark needs to save the jar to ./spark/work/appid each 
time a job get started?
Ey-Chih Chow 

Date: Sun, 8 Feb 2015 20:09:32 -0800
Subject: Re: no space left at worker node
From: 2dot7kel...@gmail.com
To: eyc...@hotmail.com
CC: gen.tan...@gmail.com; user@spark.apache.org

I guess you may set the parameters below to clean the directories:
spark.worker.cleanup.enabledspark.worker.cleanup.intervalspark.worker.cleanup.appDataTtl
They are described here: 
http://spark.apache.org/docs/1.2.0/spark-standalone.html
Kelvin
On Sun, Feb 8, 2015 at 5:15 PM, ey-chih chow  wrote:



I found the problem is, for each application, the Spark worker node saves the 
corresponding std output and std err under ./spark/work/appid, where appid is 
the id of the application.  If I ran several applications in a row, it will out 
of space.  In my case, the disk usage under ./spark/work/ is as follows:
1689784 ./app-20150208203033-0002/01689788  ./app-20150208203033-000240324  
./driver-20150208180505-00011691400 ./app-20150208180509-0001/01691404  
./app-20150208180509-000140316  ./driver-20150208203030-000240320   
./driver-20150208173156-1649876 ./app-20150208173200-/01649880  
./app-20150208173200-5152036.
Any suggestion how to resolve it?  Thanks.
Ey-Chih ChowFrom: eyc...@hotmail.com
To: gen.tan...@gmail.com
CC: user@spark.apache.org
Subject: RE: no space left at worker node
Date: Sun, 8 Feb 2015 15:25:43 -0800




By this way, the input and output paths of the job are all in s3.  I did not 
use paths of hdfs as input or output.
Best regards,
Ey-Chih Chow

From: eyc...@hotmail.com
To: gen.tan...@gmail.com
CC: user@spark.apache.org
Subject: RE: no space left at worker node
Date: Sun, 8 Feb 2015 14:57:15 -0800




Hi Gen,
Thanks.  I save my logs in a file under /var/log.  This is the only place to 
save data.  Will the problem go away if I use a better machine?
Best regards,
Ey-Chih Chow

Date: Sun, 8 Feb 2015 23:32:27 +0100
Subject: Re: no space left at worker node
From: gen.tan...@gmail.com
To: eyc...@hotmail.com
CC: user@spark.apache.org

Hi,
I am sorry that I made a mistake. r3.large has only one SSD which has been 
mounted in /mnt. Therefore this is no /dev/sdc.In fact, the problem is that 
there is no space in the under / directory. So you should check whether your 
application write data under this directory(for instance, save file in 
file:///). 
If not, you can use watch du -sh to during the running time to figure out which 
directory is expanding. Normally, only /mnt directory which is supported by SSD 
is expanding significantly, because the data of hdfs is saved here. Then you 
can find the directory which caused no space problem and find out the specific 
reason.
CheersGen


On Sun, Feb 8, 2015 at 10:45 PM, ey-chih chow  wrote:



Thanks Gen.  How can I check if /dev/sdc is well mounted or not?  In general, 
the problem shows up when I submit the second or third job.  The first job I 
submit most likely will succeed.
Ey-Chih Chow

Date: Sun, 8 Feb 2015 18:18:03 +0100
Subject: Re: no space left at worker node
From: gen.tan...@gmail.com
To: eyc...@hotma

Re: no space left at worker node

2015-02-08 Thread Kelvin Chu
Maybe, try with "local:" under the heading of Advanced Dependency
Management here:
https://spark.apache.org/docs/1.1.0/submitting-applications.html

It seems this is what you want. Hope this help.

Kelvin

On Sun, Feb 8, 2015 at 9:13 PM, ey-chih chow  wrote:

> Is there any way we can disable Spark copying the jar file to the
> corresponding directory.  I have a fat jar and is already copied to worker
> nodes using the command copydir.  Why Spark needs to save the jar to
> ./spark/work/appid each time a job get started?
>
> Ey-Chih Chow
>
> --
> Date: Sun, 8 Feb 2015 20:09:32 -0800
> Subject: Re: no space left at worker node
> From: 2dot7kel...@gmail.com
> To: eyc...@hotmail.com
> CC: gen.tan...@gmail.com; user@spark.apache.org
>
>
> I guess you may set the parameters below to clean the directories:
>
> spark.worker.cleanup.enabled
> spark.worker.cleanup.interval
> spark.worker.cleanup.appDataTtl
>
> They are described here:
> http://spark.apache.org/docs/1.2.0/spark-standalone.html
>
> Kelvin
>
> On Sun, Feb 8, 2015 at 5:15 PM, ey-chih chow  wrote:
>
> I found the problem is, for each application, the Spark worker node saves
> the corresponding std output and std err under ./spark/work/appid, where
> appid is the id of the application.  If I ran several applications in a
> row, it will out of space.  In my case, the disk usage under ./spark/work/
> is as follows:
>
> 1689784 ./app-20150208203033-0002/0
> 1689788 ./app-20150208203033-0002
> 40324 ./driver-20150208180505-0001
> 1691400 ./app-20150208180509-0001/0
> 1691404 ./app-20150208180509-0001
> 40316 ./driver-20150208203030-0002
> 40320 ./driver-20150208173156-
> 1649876 ./app-20150208173200-/0
> 1649880 ./app-20150208173200-
> 5152036 .
>
> Any suggestion how to resolve it?  Thanks.
>
> Ey-Chih Chow
> --
> From: eyc...@hotmail.com
> To: gen.tan...@gmail.com
> CC: user@spark.apache.org
> Subject: RE: no space left at worker node
> Date: Sun, 8 Feb 2015 15:25:43 -0800
>
>
> By this way, the input and output paths of the job are all in s3.  I did
> not use paths of hdfs as input or output.
>
> Best regards,
>
> Ey-Chih Chow
>
> --
> From: eyc...@hotmail.com
> To: gen.tan...@gmail.com
> CC: user@spark.apache.org
> Subject: RE: no space left at worker node
> Date: Sun, 8 Feb 2015 14:57:15 -0800
>
> Hi Gen,
>
> Thanks.  I save my logs in a file under /var/log.  This is the only place
> to save data.  Will the problem go away if I use a better machine?
>
> Best regards,
>
> Ey-Chih Chow
>
> --
> Date: Sun, 8 Feb 2015 23:32:27 +0100
> Subject: Re: no space left at worker node
> From: gen.tan...@gmail.com
> To: eyc...@hotmail.com
> CC: user@spark.apache.org
>
> Hi,
>
> I am sorry that I made a mistake. r3.large has only one SSD which has been
> mounted in /mnt. Therefore this is no /dev/sdc.
> In fact, the problem is that there is no space in the under / directory.
> So you should check whether your application write data under this
> directory(for instance, save file in file:///).
>
> If not, you can use watch du -sh to during the running time to figure out
> which directory is expanding. Normally, only /mnt directory which is
> supported by SSD is expanding significantly, because the data of hdfs is
> saved here. Then you can find the directory which caused no space problem
> and find out the specific reason.
>
> Cheers
> Gen
>
>
>
> On Sun, Feb 8, 2015 at 10:45 PM, ey-chih chow  wrote:
>
> Thanks Gen.  How can I check if /dev/sdc is well mounted or not?  In
> general, the problem shows up when I submit the second or third job.  The
> first job I submit most likely will succeed.
>
> Ey-Chih Chow
>
> --
> Date: Sun, 8 Feb 2015 18:18:03 +0100
>
> Subject: Re: no space left at worker node
> From: gen.tan...@gmail.com
> To: eyc...@hotmail.com
> CC: user@spark.apache.org
>
> Hi,
>
> In fact, /dev/sdb is /dev/xvdb. It seems that there is no problem about
> double mount. However, there is no information about /mnt2. You should
> check whether /dev/sdc is well mounted or not.
> The reply of Micheal is good solution about this type of problem. You can
> check his site.
>
> Cheers
> Gen
>
>
> On Sun, Feb 8, 2015 at 5:53 PM, ey-chih chow  wrote:
>
> Gen,
>
> Thanks for your information.  The content of /etc/fstab at the worker node
> (r3.large) is:
>
> #
> LABEL=/ /   ext4defaults,noatime  1   1
> tmpfs   /dev/shmtmpfs   defaults0   0
> devpts  /de

RE: no space left at worker node

2015-02-08 Thread ey-chih chow
Is there any way we can disable Spark copying the jar file to the corresponding 
directory.  I have a fat jar and is already copied to worker nodes using the 
command copydir.  Why Spark needs to save the jar to ./spark/work/appid each 
time a job get started?
Ey-Chih Chow 

Date: Sun, 8 Feb 2015 20:09:32 -0800
Subject: Re: no space left at worker node
From: 2dot7kel...@gmail.com
To: eyc...@hotmail.com
CC: gen.tan...@gmail.com; user@spark.apache.org

I guess you may set the parameters below to clean the directories:
spark.worker.cleanup.enabledspark.worker.cleanup.intervalspark.worker.cleanup.appDataTtl
They are described here: 
http://spark.apache.org/docs/1.2.0/spark-standalone.html
Kelvin
On Sun, Feb 8, 2015 at 5:15 PM, ey-chih chow  wrote:



I found the problem is, for each application, the Spark worker node saves the 
corresponding std output and std err under ./spark/work/appid, where appid is 
the id of the application.  If I ran several applications in a row, it will out 
of space.  In my case, the disk usage under ./spark/work/ is as follows:
1689784 ./app-20150208203033-0002/01689788  ./app-20150208203033-000240324  
./driver-20150208180505-00011691400 ./app-20150208180509-0001/01691404  
./app-20150208180509-000140316  ./driver-20150208203030-000240320   
./driver-20150208173156-1649876 ./app-20150208173200-/01649880  
./app-20150208173200-5152036.
Any suggestion how to resolve it?  Thanks.
Ey-Chih ChowFrom: eyc...@hotmail.com
To: gen.tan...@gmail.com
CC: user@spark.apache.org
Subject: RE: no space left at worker node
Date: Sun, 8 Feb 2015 15:25:43 -0800




By this way, the input and output paths of the job are all in s3.  I did not 
use paths of hdfs as input or output.
Best regards,
Ey-Chih Chow

From: eyc...@hotmail.com
To: gen.tan...@gmail.com
CC: user@spark.apache.org
Subject: RE: no space left at worker node
Date: Sun, 8 Feb 2015 14:57:15 -0800




Hi Gen,
Thanks.  I save my logs in a file under /var/log.  This is the only place to 
save data.  Will the problem go away if I use a better machine?
Best regards,
Ey-Chih Chow

Date: Sun, 8 Feb 2015 23:32:27 +0100
Subject: Re: no space left at worker node
From: gen.tan...@gmail.com
To: eyc...@hotmail.com
CC: user@spark.apache.org

Hi,
I am sorry that I made a mistake. r3.large has only one SSD which has been 
mounted in /mnt. Therefore this is no /dev/sdc.In fact, the problem is that 
there is no space in the under / directory. So you should check whether your 
application write data under this directory(for instance, save file in 
file:///). 
If not, you can use watch du -sh to during the running time to figure out which 
directory is expanding. Normally, only /mnt directory which is supported by SSD 
is expanding significantly, because the data of hdfs is saved here. Then you 
can find the directory which caused no space problem and find out the specific 
reason.
CheersGen


On Sun, Feb 8, 2015 at 10:45 PM, ey-chih chow  wrote:



Thanks Gen.  How can I check if /dev/sdc is well mounted or not?  In general, 
the problem shows up when I submit the second or third job.  The first job I 
submit most likely will succeed.
Ey-Chih Chow

Date: Sun, 8 Feb 2015 18:18:03 +0100
Subject: Re: no space left at worker node
From: gen.tan...@gmail.com
To: eyc...@hotmail.com
CC: user@spark.apache.org

Hi,
In fact, /dev/sdb is /dev/xvdb. It seems that there is no problem about double 
mount. However, there is no information about /mnt2. You should check whether 
/dev/sdc is well mounted or not.The reply of Micheal is good solution about 
this type of problem. You can check his site.
CheersGen

On Sun, Feb 8, 2015 at 5:53 PM, ey-chih chow  wrote:



Gen,
Thanks for your information.  The content of /etc/fstab at the worker node 
(r3.large) is:
#LABEL=/ /   ext4defaults,noatime  1   1tmpfs   /dev/shm
tmpfs   defaults0   0devpts  /dev/ptsdevpts  gid=5,mode=620  0  
 0sysfs   /syssysfs   defaults0   0proc/proc   
procdefaults0   0/dev/sdb/mntauto
defaults,noatime,nodiratime,comment=cloudconfig 0   0/dev/sdc/mnt2  
 autodefaults,noatime,nodiratime,comment=cloudconfig 0   0
There is no entry of /dev/xvdb.
 Ey-Chih Chow
Date: Sun, 8 Feb 2015 12:09:37 +0100
Subject: Re: no space left at worker node
From: gen.tan...@gmail.com
To: eyc...@hotmail.com
CC: user@spark.apache.org

Hi,
I fact, I met this problem before. it is a bug of AWS. Which type of machine do 
you use?
If I guess well, you can check the file /etc/fstab. There would be a double 
mount of /dev/xvdb.If yes, you should1. stop hdfs2. umount /dev/xvdb at / 3. 
restart hdfs
Hope this could be helpful.CheersGen


On Sun, Feb 8, 2015 at 8:16 AM, ey-chih chow  wrote:
Hi,



I submitted a spark job to an ec2 cluster, using spark-submit.  At a worker

node, there is an exception of 'no space left on device' 

Re: no space left at worker node

2015-02-08 Thread Kelvin Chu
I guess you may set the parameters below to clean the directories:

spark.worker.cleanup.enabled
spark.worker.cleanup.interval
spark.worker.cleanup.appDataTtl

They are described here:
http://spark.apache.org/docs/1.2.0/spark-standalone.html

Kelvin

On Sun, Feb 8, 2015 at 5:15 PM, ey-chih chow  wrote:

> I found the problem is, for each application, the Spark worker node saves
> the corresponding std output and std err under ./spark/work/appid, where
> appid is the id of the application.  If I ran several applications in a
> row, it will out of space.  In my case, the disk usage under ./spark/work/
> is as follows:
>
> 1689784 ./app-20150208203033-0002/0
> 1689788 ./app-20150208203033-0002
> 40324 ./driver-20150208180505-0001
> 1691400 ./app-20150208180509-0001/0
> 1691404 ./app-20150208180509-0001
> 40316 ./driver-20150208203030-0002
> 40320 ./driver-20150208173156-
> 1649876 ./app-20150208173200-/0
> 1649880 ./app-20150208173200-
> 5152036 .
>
> Any suggestion how to resolve it?  Thanks.
>
> Ey-Chih Chow
> --
> From: eyc...@hotmail.com
> To: gen.tan...@gmail.com
> CC: user@spark.apache.org
> Subject: RE: no space left at worker node
> Date: Sun, 8 Feb 2015 15:25:43 -0800
>
>
> By this way, the input and output paths of the job are all in s3.  I did
> not use paths of hdfs as input or output.
>
> Best regards,
>
> Ey-Chih Chow
>
> ----------
> From: eyc...@hotmail.com
> To: gen.tan...@gmail.com
> CC: user@spark.apache.org
> Subject: RE: no space left at worker node
> Date: Sun, 8 Feb 2015 14:57:15 -0800
>
> Hi Gen,
>
> Thanks.  I save my logs in a file under /var/log.  This is the only place
> to save data.  Will the problem go away if I use a better machine?
>
> Best regards,
>
> Ey-Chih Chow
>
> --
> Date: Sun, 8 Feb 2015 23:32:27 +0100
> Subject: Re: no space left at worker node
> From: gen.tan...@gmail.com
> To: eyc...@hotmail.com
> CC: user@spark.apache.org
>
> Hi,
>
> I am sorry that I made a mistake. r3.large has only one SSD which has been
> mounted in /mnt. Therefore this is no /dev/sdc.
> In fact, the problem is that there is no space in the under / directory.
> So you should check whether your application write data under this
> directory(for instance, save file in file:///).
>
> If not, you can use watch du -sh to during the running time to figure out
> which directory is expanding. Normally, only /mnt directory which is
> supported by SSD is expanding significantly, because the data of hdfs is
> saved here. Then you can find the directory which caused no space problem
> and find out the specific reason.
>
> Cheers
> Gen
>
>
>
> On Sun, Feb 8, 2015 at 10:45 PM, ey-chih chow  wrote:
>
> Thanks Gen.  How can I check if /dev/sdc is well mounted or not?  In
> general, the problem shows up when I submit the second or third job.  The
> first job I submit most likely will succeed.
>
> Ey-Chih Chow
>
> --
> Date: Sun, 8 Feb 2015 18:18:03 +0100
>
> Subject: Re: no space left at worker node
> From: gen.tan...@gmail.com
> To: eyc...@hotmail.com
> CC: user@spark.apache.org
>
> Hi,
>
> In fact, /dev/sdb is /dev/xvdb. It seems that there is no problem about
> double mount. However, there is no information about /mnt2. You should
> check whether /dev/sdc is well mounted or not.
> The reply of Micheal is good solution about this type of problem. You can
> check his site.
>
> Cheers
> Gen
>
>
> On Sun, Feb 8, 2015 at 5:53 PM, ey-chih chow  wrote:
>
> Gen,
>
> Thanks for your information.  The content of /etc/fstab at the worker node
> (r3.large) is:
>
> #
> LABEL=/ /   ext4defaults,noatime  1   1
> tmpfs   /dev/shmtmpfs   defaults0   0
> devpts  /dev/ptsdevpts  gid=5,mode=620  0   0
> sysfs   /syssysfs   defaults0   0
> proc/proc   proc    defaults    0   0
> /dev/sdb/mntauto
>  defaults,noatime,nodiratime,comment=cloudconfig 0   0
> /dev/sdc/mnt2   auto
>  defaults,noatime,nodiratime,comment=cloudconfig 0   0
>
> There is no entry of /dev/xvdb.
>
>  Ey-Chih Chow
>
> --
> Date: Sun, 8 Feb 2015 12:09:37 +0100
> Subject: Re: no space left at worker node
> From: gen.tan...@gmail.com
> To: eyc...@hotmail.com
> CC: user@spark.apache.org
>
>
> Hi,
>
> I fact, I met this problem before. it is a bug of AWS. Which type of
> machine do you use?
>
> If I guess well, you can check the file /etc/fstab. There would be a
> double mount of /dev/

RE: no space left at worker node

2015-02-08 Thread ey-chih chow
I found the problem is, for each application, the Spark worker node saves the 
corresponding std output and std err under ./spark/work/appid, where appid is 
the id of the application.  If I ran several applications in a row, it will out 
of space.  In my case, the disk usage under ./spark/work/ is as follows:
1689784 ./app-20150208203033-0002/01689788  ./app-20150208203033-000240324  
./driver-20150208180505-00011691400 ./app-20150208180509-0001/01691404  
./app-20150208180509-000140316  ./driver-20150208203030-000240320   
./driver-20150208173156-1649876 ./app-20150208173200-/01649880  
./app-20150208173200-5152036.
Any suggestion how to resolve it?  Thanks.
Ey-Chih ChowFrom: eyc...@hotmail.com
To: gen.tan...@gmail.com
CC: user@spark.apache.org
Subject: RE: no space left at worker node
Date: Sun, 8 Feb 2015 15:25:43 -0800




By this way, the input and output paths of the job are all in s3.  I did not 
use paths of hdfs as input or output.
Best regards,
Ey-Chih Chow

From: eyc...@hotmail.com
To: gen.tan...@gmail.com
CC: user@spark.apache.org
Subject: RE: no space left at worker node
Date: Sun, 8 Feb 2015 14:57:15 -0800




Hi Gen,
Thanks.  I save my logs in a file under /var/log.  This is the only place to 
save data.  Will the problem go away if I use a better machine?
Best regards,
Ey-Chih Chow

Date: Sun, 8 Feb 2015 23:32:27 +0100
Subject: Re: no space left at worker node
From: gen.tan...@gmail.com
To: eyc...@hotmail.com
CC: user@spark.apache.org

Hi,
I am sorry that I made a mistake. r3.large has only one SSD which has been 
mounted in /mnt. Therefore this is no /dev/sdc.In fact, the problem is that 
there is no space in the under / directory. So you should check whether your 
application write data under this directory(for instance, save file in 
file:///). 
If not, you can use watch du -sh to during the running time to figure out which 
directory is expanding. Normally, only /mnt directory which is supported by SSD 
is expanding significantly, because the data of hdfs is saved here. Then you 
can find the directory which caused no space problem and find out the specific 
reason.
CheersGen


On Sun, Feb 8, 2015 at 10:45 PM, ey-chih chow  wrote:



Thanks Gen.  How can I check if /dev/sdc is well mounted or not?  In general, 
the problem shows up when I submit the second or third job.  The first job I 
submit most likely will succeed.
Ey-Chih Chow

Date: Sun, 8 Feb 2015 18:18:03 +0100
Subject: Re: no space left at worker node
From: gen.tan...@gmail.com
To: eyc...@hotmail.com
CC: user@spark.apache.org

Hi,
In fact, /dev/sdb is /dev/xvdb. It seems that there is no problem about double 
mount. However, there is no information about /mnt2. You should check whether 
/dev/sdc is well mounted or not.The reply of Micheal is good solution about 
this type of problem. You can check his site.
CheersGen

On Sun, Feb 8, 2015 at 5:53 PM, ey-chih chow  wrote:



Gen,
Thanks for your information.  The content of /etc/fstab at the worker node 
(r3.large) is:
#LABEL=/ /   ext4defaults,noatime  1   1tmpfs   /dev/shm
tmpfs   defaults0   0devpts  /dev/ptsdevpts  gid=5,mode=620  0  
 0sysfs   /syssysfs   defaults0   0proc/proc   
procdefaults0   0/dev/sdb/mntauto
defaults,noatime,nodiratime,comment=cloudconfig 0   0/dev/sdc/mnt2  
 autodefaults,noatime,nodiratime,comment=cloudconfig 0   0
There is no entry of /dev/xvdb.
 Ey-Chih Chow
Date: Sun, 8 Feb 2015 12:09:37 +0100
Subject: Re: no space left at worker node
From: gen.tan...@gmail.com
To: eyc...@hotmail.com
CC: user@spark.apache.org

Hi,
I fact, I met this problem before. it is a bug of AWS. Which type of machine do 
you use?
If I guess well, you can check the file /etc/fstab. There would be a double 
mount of /dev/xvdb.If yes, you should1. stop hdfs2. umount /dev/xvdb at / 3. 
restart hdfs
Hope this could be helpful.CheersGen


On Sun, Feb 8, 2015 at 8:16 AM, ey-chih chow  wrote:
Hi,



I submitted a spark job to an ec2 cluster, using spark-submit.  At a worker

node, there is an exception of 'no space left on device' as follows.



==

15/02/08 01:53:38 ERROR logging.FileAppender: Error writing stream to file

/root/spark/work/app-20150208014557-0003/0/stdout

java.io.IOException: No space left on device

at java.io.FileOutputStream.writeBytes(Native Method)

at java.io.FileOutputStream.write(FileOutputStream.java:345)

at

org.apache.spark.util.logging.FileAppender.appendToFile(FileAppender.scala:92)

at

org.apache.spark.util.logging.FileAppender.appendStreamToFile(FileAppender.scala:72)

at

org.apache.spark.util.logging.FileAppender$$anon$1$$anonfun$run$1.apply$mcV$sp(FileAppender.scala:39)

at

org.apache.spark.util.logging.FileAppender$$anon$1$$anonfun$run$1.apply(FileAppender.scala:39)


RE: no space left at worker node

2015-02-08 Thread ey-chih chow
By this way, the input and output paths of the job are all in s3.  I did not 
use paths of hdfs as input or output.
Best regards,
Ey-Chih Chow

From: eyc...@hotmail.com
To: gen.tan...@gmail.com
CC: user@spark.apache.org
Subject: RE: no space left at worker node
Date: Sun, 8 Feb 2015 14:57:15 -0800




Hi Gen,
Thanks.  I save my logs in a file under /var/log.  This is the only place to 
save data.  Will the problem go away if I use a better machine?
Best regards,
Ey-Chih Chow

Date: Sun, 8 Feb 2015 23:32:27 +0100
Subject: Re: no space left at worker node
From: gen.tan...@gmail.com
To: eyc...@hotmail.com
CC: user@spark.apache.org

Hi,
I am sorry that I made a mistake. r3.large has only one SSD which has been 
mounted in /mnt. Therefore this is no /dev/sdc.In fact, the problem is that 
there is no space in the under / directory. So you should check whether your 
application write data under this directory(for instance, save file in 
file:///). 
If not, you can use watch du -sh to during the running time to figure out which 
directory is expanding. Normally, only /mnt directory which is supported by SSD 
is expanding significantly, because the data of hdfs is saved here. Then you 
can find the directory which caused no space problem and find out the specific 
reason.
CheersGen


On Sun, Feb 8, 2015 at 10:45 PM, ey-chih chow  wrote:



Thanks Gen.  How can I check if /dev/sdc is well mounted or not?  In general, 
the problem shows up when I submit the second or third job.  The first job I 
submit most likely will succeed.
Ey-Chih Chow

Date: Sun, 8 Feb 2015 18:18:03 +0100
Subject: Re: no space left at worker node
From: gen.tan...@gmail.com
To: eyc...@hotmail.com
CC: user@spark.apache.org

Hi,
In fact, /dev/sdb is /dev/xvdb. It seems that there is no problem about double 
mount. However, there is no information about /mnt2. You should check whether 
/dev/sdc is well mounted or not.The reply of Micheal is good solution about 
this type of problem. You can check his site.
CheersGen

On Sun, Feb 8, 2015 at 5:53 PM, ey-chih chow  wrote:



Gen,
Thanks for your information.  The content of /etc/fstab at the worker node 
(r3.large) is:
#LABEL=/ /   ext4defaults,noatime  1   1tmpfs   /dev/shm
tmpfs   defaults0   0devpts  /dev/ptsdevpts  gid=5,mode=620  0  
 0sysfs   /syssysfs   defaults0   0proc/proc   
procdefaults0   0/dev/sdb/mntauto
defaults,noatime,nodiratime,comment=cloudconfig 0   0/dev/sdc/mnt2  
 autodefaults,noatime,nodiratime,comment=cloudconfig 0   0
There is no entry of /dev/xvdb.
 Ey-Chih Chow
Date: Sun, 8 Feb 2015 12:09:37 +0100
Subject: Re: no space left at worker node
From: gen.tan...@gmail.com
To: eyc...@hotmail.com
CC: user@spark.apache.org

Hi,
I fact, I met this problem before. it is a bug of AWS. Which type of machine do 
you use?
If I guess well, you can check the file /etc/fstab. There would be a double 
mount of /dev/xvdb.If yes, you should1. stop hdfs2. umount /dev/xvdb at / 3. 
restart hdfs
Hope this could be helpful.CheersGen


On Sun, Feb 8, 2015 at 8:16 AM, ey-chih chow  wrote:
Hi,



I submitted a spark job to an ec2 cluster, using spark-submit.  At a worker

node, there is an exception of 'no space left on device' as follows.



==

15/02/08 01:53:38 ERROR logging.FileAppender: Error writing stream to file

/root/spark/work/app-20150208014557-0003/0/stdout

java.io.IOException: No space left on device

at java.io.FileOutputStream.writeBytes(Native Method)

at java.io.FileOutputStream.write(FileOutputStream.java:345)

at

org.apache.spark.util.logging.FileAppender.appendToFile(FileAppender.scala:92)

at

org.apache.spark.util.logging.FileAppender.appendStreamToFile(FileAppender.scala:72)

at

org.apache.spark.util.logging.FileAppender$$anon$1$$anonfun$run$1.apply$mcV$sp(FileAppender.scala:39)

at

org.apache.spark.util.logging.FileAppender$$anon$1$$anonfun$run$1.apply(FileAppender.scala:39)

at

org.apache.spark.util.logging.FileAppender$$anon$1$$anonfun$run$1.apply(FileAppender.scala:39)

at

org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:1311)

at

org.apache.spark.util.logging.FileAppender$$anon$1.run(FileAppender.scala:38)

===



The command df showed the following information at the worker node:



Filesystem   1K-blocks  Used Available Use% Mounted on

/dev/xvda1 8256920   8256456 0 100% /

tmpfs  7752012 0   7752012   0% /dev/shm

/dev/xvdb 30963708   1729652  27661192   6% /mnt



Does anybody know how to fix this?  Thanks.





Ey-Chih Chow







--

View this message in context: 
http://apache-spark-user-list.1001560.n3.nabble.com/no-space-left-at-worker-node-tp21545.html

Sent from the Apache Spark User List ma

RE: no space left at worker node

2015-02-08 Thread ey-chih chow
Hi Gen,
Thanks.  I save my logs in a file under /var/log.  This is the only place to 
save data.  Will the problem go away if I use a better machine?
Best regards,
Ey-Chih Chow

Date: Sun, 8 Feb 2015 23:32:27 +0100
Subject: Re: no space left at worker node
From: gen.tan...@gmail.com
To: eyc...@hotmail.com
CC: user@spark.apache.org

Hi,
I am sorry that I made a mistake. r3.large has only one SSD which has been 
mounted in /mnt. Therefore this is no /dev/sdc.In fact, the problem is that 
there is no space in the under / directory. So you should check whether your 
application write data under this directory(for instance, save file in 
file:///). 
If not, you can use watch du -sh to during the running time to figure out which 
directory is expanding. Normally, only /mnt directory which is supported by SSD 
is expanding significantly, because the data of hdfs is saved here. Then you 
can find the directory which caused no space problem and find out the specific 
reason.
CheersGen


On Sun, Feb 8, 2015 at 10:45 PM, ey-chih chow  wrote:



Thanks Gen.  How can I check if /dev/sdc is well mounted or not?  In general, 
the problem shows up when I submit the second or third job.  The first job I 
submit most likely will succeed.
Ey-Chih Chow

Date: Sun, 8 Feb 2015 18:18:03 +0100
Subject: Re: no space left at worker node
From: gen.tan...@gmail.com
To: eyc...@hotmail.com
CC: user@spark.apache.org

Hi,
In fact, /dev/sdb is /dev/xvdb. It seems that there is no problem about double 
mount. However, there is no information about /mnt2. You should check whether 
/dev/sdc is well mounted or not.The reply of Micheal is good solution about 
this type of problem. You can check his site.
CheersGen

On Sun, Feb 8, 2015 at 5:53 PM, ey-chih chow  wrote:



Gen,
Thanks for your information.  The content of /etc/fstab at the worker node 
(r3.large) is:
#LABEL=/ /   ext4defaults,noatime  1   1tmpfs   /dev/shm
tmpfs   defaults0   0devpts  /dev/ptsdevpts  gid=5,mode=620  0  
 0sysfs   /syssysfs   defaults0   0proc/proc   
procdefaults0   0/dev/sdb/mntauto
defaults,noatime,nodiratime,comment=cloudconfig 0   0/dev/sdc/mnt2  
 autodefaults,noatime,nodiratime,comment=cloudconfig 0   0
There is no entry of /dev/xvdb.
 Ey-Chih Chow
Date: Sun, 8 Feb 2015 12:09:37 +0100
Subject: Re: no space left at worker node
From: gen.tan...@gmail.com
To: eyc...@hotmail.com
CC: user@spark.apache.org

Hi,
I fact, I met this problem before. it is a bug of AWS. Which type of machine do 
you use?
If I guess well, you can check the file /etc/fstab. There would be a double 
mount of /dev/xvdb.If yes, you should1. stop hdfs2. umount /dev/xvdb at / 3. 
restart hdfs
Hope this could be helpful.CheersGen


On Sun, Feb 8, 2015 at 8:16 AM, ey-chih chow  wrote:
Hi,



I submitted a spark job to an ec2 cluster, using spark-submit.  At a worker

node, there is an exception of 'no space left on device' as follows.



==

15/02/08 01:53:38 ERROR logging.FileAppender: Error writing stream to file

/root/spark/work/app-20150208014557-0003/0/stdout

java.io.IOException: No space left on device

at java.io.FileOutputStream.writeBytes(Native Method)

at java.io.FileOutputStream.write(FileOutputStream.java:345)

at

org.apache.spark.util.logging.FileAppender.appendToFile(FileAppender.scala:92)

at

org.apache.spark.util.logging.FileAppender.appendStreamToFile(FileAppender.scala:72)

at

org.apache.spark.util.logging.FileAppender$$anon$1$$anonfun$run$1.apply$mcV$sp(FileAppender.scala:39)

at

org.apache.spark.util.logging.FileAppender$$anon$1$$anonfun$run$1.apply(FileAppender.scala:39)

at

org.apache.spark.util.logging.FileAppender$$anon$1$$anonfun$run$1.apply(FileAppender.scala:39)

at

org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:1311)

at

org.apache.spark.util.logging.FileAppender$$anon$1.run(FileAppender.scala:38)

===



The command df showed the following information at the worker node:



Filesystem   1K-blocks  Used Available Use% Mounted on

/dev/xvda1 8256920   8256456 0 100% /

tmpfs  7752012 0   7752012   0% /dev/shm

/dev/xvdb 30963708   1729652  27661192   6% /mnt



Does anybody know how to fix this?  Thanks.





Ey-Chih Chow







--

View this message in context: 
http://apache-spark-user-list.1001560.n3.nabble.com/no-space-left-at-worker-node-tp21545.html

Sent from the Apache Spark User List mailing list archive at Nabble.com.



-

To unsubscribe, e-mail: user-unsubscr...@spark.apache.org

For additional commands, e-mail: user-h...@spark.apache.org




  

  

  

Re: no space left at worker node

2015-02-08 Thread gen tang
Hi,

I am sorry that I made a mistake. r3.large has only one SSD which has been
mounted in /mnt. Therefore this is no /dev/sdc.
In fact, the problem is that there is no space in the under / directory. So
you should check whether your application write data under this
directory(for instance, save file in file:///).

If not, you can use watch du -sh to during the running time to figure out
which directory is expanding. Normally, only /mnt directory which is
supported by SSD is expanding significantly, because the data of hdfs is
saved here. Then you can find the directory which caused no space problem
and find out the specific reason.

Cheers
Gen



On Sun, Feb 8, 2015 at 10:45 PM, ey-chih chow  wrote:

> Thanks Gen.  How can I check if /dev/sdc is well mounted or not?  In
> general, the problem shows up when I submit the second or third job.  The
> first job I submit most likely will succeed.
>
> Ey-Chih Chow
>
> --
> Date: Sun, 8 Feb 2015 18:18:03 +0100
>
> Subject: Re: no space left at worker node
> From: gen.tan...@gmail.com
> To: eyc...@hotmail.com
> CC: user@spark.apache.org
>
> Hi,
>
> In fact, /dev/sdb is /dev/xvdb. It seems that there is no problem about
> double mount. However, there is no information about /mnt2. You should
> check whether /dev/sdc is well mounted or not.
> The reply of Micheal is good solution about this type of problem. You can
> check his site.
>
> Cheers
> Gen
>
>
> On Sun, Feb 8, 2015 at 5:53 PM, ey-chih chow  wrote:
>
> Gen,
>
> Thanks for your information.  The content of /etc/fstab at the worker node
> (r3.large) is:
>
> #
> LABEL=/ /   ext4defaults,noatime  1   1
> tmpfs   /dev/shmtmpfs   defaults0   0
> devpts  /dev/ptsdevpts  gid=5,mode=620  0   0
> sysfs   /syssysfs   defaults0   0
> proc/proc   procdefaults0   0
> /dev/sdb/mntauto
>  defaults,noatime,nodiratime,comment=cloudconfig 0   0
> /dev/sdc/mnt2   auto
>  defaults,noatime,nodiratime,comment=cloudconfig 0   0
>
> There is no entry of /dev/xvdb.
>
>  Ey-Chih Chow
>
> --
> Date: Sun, 8 Feb 2015 12:09:37 +0100
> Subject: Re: no space left at worker node
> From: gen.tan...@gmail.com
> To: eyc...@hotmail.com
> CC: user@spark.apache.org
>
>
> Hi,
>
> I fact, I met this problem before. it is a bug of AWS. Which type of
> machine do you use?
>
> If I guess well, you can check the file /etc/fstab. There would be a
> double mount of /dev/xvdb.
> If yes, you should
> 1. stop hdfs
> 2. umount /dev/xvdb at /
> 3. restart hdfs
>
> Hope this could be helpful.
> Cheers
> Gen
>
>
>
> On Sun, Feb 8, 2015 at 8:16 AM, ey-chih chow  wrote:
>
> Hi,
>
> I submitted a spark job to an ec2 cluster, using spark-submit.  At a worker
> node, there is an exception of 'no space left on device' as follows.
>
> ==
> 15/02/08 01:53:38 ERROR logging.FileAppender: Error writing stream to file
> /root/spark/work/app-20150208014557-0003/0/stdout
> java.io.IOException: No space left on device
> at java.io.FileOutputStream.writeBytes(Native Method)
> at java.io.FileOutputStream.write(FileOutputStream.java:345)
> at
>
> org.apache.spark.util.logging.FileAppender.appendToFile(FileAppender.scala:92)
> at
>
> org.apache.spark.util.logging.FileAppender.appendStreamToFile(FileAppender.scala:72)
> at
>
> org.apache.spark.util.logging.FileAppender$$anon$1$$anonfun$run$1.apply$mcV$sp(FileAppender.scala:39)
> at
>
> org.apache.spark.util.logging.FileAppender$$anon$1$$anonfun$run$1.apply(FileAppender.scala:39)
> at
>
> org.apache.spark.util.logging.FileAppender$$anon$1$$anonfun$run$1.apply(FileAppender.scala:39)
> at
> org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:1311)
> at
>
> org.apache.spark.util.logging.FileAppender$$anon$1.run(FileAppender.scala:38)
> ===
>
> The command df showed the following information at the worker node:
>
> Filesystem   1K-blocks  Used Available Use% Mounted on
> /dev/xvda1 8256920   8256456 0 100% /
> tmpfs  7752012 0   7752012   0% /dev/shm
> /dev/xvdb 30963708   1729652  27661192   6% /mnt
>
> Does anybody know how to fix this?  Thanks.
>
>
> Ey-Chih Chow
>
>
>
> --
> View this message in context:
> http://apache-spark-user-list.1001560.n3.nabble.com/no-space-left-at-worker-node-tp21545.html
> Sent from the Apache Spark User List mailing list archive at Nabble.com.
>
> -
> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
> For additional commands, e-mail: user-h...@spark.apache.org
>
>
>
>


RE: no space left at worker node

2015-02-08 Thread ey-chih chow
Thanks Michael.  I didn't edit core-site.xml.  We use the default one.  I only 
saw hdaoop.tmp.dir in core-site.xml, pointing to /mnt/ephemeral-hdfs.  How can 
I edit the config file?
Best regards,
Ey-Chih

Date: Sun, 8 Feb 2015 16:51:32 +
From: m_albert...@yahoo.com
To: gen.tan...@gmail.com; eyc...@hotmail.com
CC: user@spark.apache.org
Subject: Re: no space left at worker node

You might want to take a look in core-site.xml, andsee what is listed as usable 
directories (hadoop.tmp.dir, fs.s3.buffer.dir).
It seems that on S3, the root disk is relatively small (8G), but the config 
files list a "mnt" directory under it.  Somehow the system doesn't balance 
between the very small space it has under the root disk and the larger disks, 
so the root disk fills up while the others are unused.
At my site, we wrote a boot script to edit these problem out of the config 
before hadoop starts.
-Mike
From: gen tang 
 To: ey-chih chow  
Cc: "user@spark.apache.org"  
 Sent: Sunday, February 8, 2015 6:09 AM
 Subject: Re: no space left at worker node
   
Hi,I fact, I met this problem before. it is a bug of AWS. Which type of machine 
do you use?If I guess well, you can check the file /etc/fstab. There would be a 
double mount of /dev/xvdb.If yes, you should1. stop hdfs2. umount /dev/xvdb at 
/ 3. restart hdfsHope this could be helpful.CheersGen

On Sun, Feb 8, 2015 at 8:16 AM, ey-chih chow  wrote:Hi,

I submitted a spark job to an ec2 cluster, using spark-submit.  At a worker
node, there is an exception of 'no space left on device' as follows.

==
15/02/08 01:53:38 ERROR logging.FileAppender: Error writing stream to file
/root/spark/work/app-20150208014557-0003/0/stdout
java.io.IOException: No space left on device
at java.io.FileOutputStream.writeBytes(Native Method)
at java.io.FileOutputStream.write(FileOutputStream.java:345)
at
org.apache.spark.util.logging.FileAppender.appendToFile(FileAppender.scala:92)
at
org.apache.spark.util.logging.FileAppender.appendStreamToFile(FileAppender.scala:72)
at
org.apache.spark.util.logging.FileAppender$$anon$1$$anonfun$run$1.apply$mcV$sp(FileAppender.scala:39)
at
org.apache.spark.util.logging.FileAppender$$anon$1$$anonfun$run$1.apply(FileAppender.scala:39)
at
org.apache.spark.util.logging.FileAppender$$anon$1$$anonfun$run$1.apply(FileAppender.scala:39)
at
org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:1311)
at
org.apache.spark.util.logging.FileAppender$$anon$1.run(FileAppender.scala:38)
===

The command df showed the following information at the worker node:

Filesystem   1K-blocks  Used Available Use% Mounted on
/dev/xvda1 8256920   8256456 0 100% /
tmpfs  7752012 0   7752012   0% /dev/shm
/dev/xvdb 30963708   1729652  27661192   6% /mnt

Does anybody know how to fix this?  Thanks.


Ey-Chih Chow



--
View this message in context: 
http://apache-spark-user-list.1001560.n3.nabble.com/no-space-left-at-worker-node-tp21545.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.

-
To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
For additional commands, e-mail: user-h...@spark.apache.org



  

RE: no space left at worker node

2015-02-08 Thread ey-chih chow
Thanks Gen.  How can I check if /dev/sdc is well mounted or not?  In general, 
the problem shows up when I submit the second or third job.  The first job I 
submit most likely will succeed.
Ey-Chih Chow

Date: Sun, 8 Feb 2015 18:18:03 +0100
Subject: Re: no space left at worker node
From: gen.tan...@gmail.com
To: eyc...@hotmail.com
CC: user@spark.apache.org

Hi,
In fact, /dev/sdb is /dev/xvdb. It seems that there is no problem about double 
mount. However, there is no information about /mnt2. You should check whether 
/dev/sdc is well mounted or not.The reply of Micheal is good solution about 
this type of problem. You can check his site.
CheersGen

On Sun, Feb 8, 2015 at 5:53 PM, ey-chih chow  wrote:



Gen,
Thanks for your information.  The content of /etc/fstab at the worker node 
(r3.large) is:
#LABEL=/ /   ext4defaults,noatime  1   1tmpfs   /dev/shm
tmpfs   defaults0   0devpts  /dev/ptsdevpts  gid=5,mode=620  0  
 0sysfs   /syssysfs   defaults0   0proc/proc   
procdefaults0   0/dev/sdb/mntauto
defaults,noatime,nodiratime,comment=cloudconfig 0   0/dev/sdc/mnt2  
 autodefaults,noatime,nodiratime,comment=cloudconfig 0   0
There is no entry of /dev/xvdb.
 Ey-Chih Chow
Date: Sun, 8 Feb 2015 12:09:37 +0100
Subject: Re: no space left at worker node
From: gen.tan...@gmail.com
To: eyc...@hotmail.com
CC: user@spark.apache.org

Hi,
I fact, I met this problem before. it is a bug of AWS. Which type of machine do 
you use?
If I guess well, you can check the file /etc/fstab. There would be a double 
mount of /dev/xvdb.If yes, you should1. stop hdfs2. umount /dev/xvdb at / 3. 
restart hdfs
Hope this could be helpful.CheersGen


On Sun, Feb 8, 2015 at 8:16 AM, ey-chih chow  wrote:
Hi,



I submitted a spark job to an ec2 cluster, using spark-submit.  At a worker

node, there is an exception of 'no space left on device' as follows.



==

15/02/08 01:53:38 ERROR logging.FileAppender: Error writing stream to file

/root/spark/work/app-20150208014557-0003/0/stdout

java.io.IOException: No space left on device

at java.io.FileOutputStream.writeBytes(Native Method)

at java.io.FileOutputStream.write(FileOutputStream.java:345)

at

org.apache.spark.util.logging.FileAppender.appendToFile(FileAppender.scala:92)

at

org.apache.spark.util.logging.FileAppender.appendStreamToFile(FileAppender.scala:72)

at

org.apache.spark.util.logging.FileAppender$$anon$1$$anonfun$run$1.apply$mcV$sp(FileAppender.scala:39)

at

org.apache.spark.util.logging.FileAppender$$anon$1$$anonfun$run$1.apply(FileAppender.scala:39)

at

org.apache.spark.util.logging.FileAppender$$anon$1$$anonfun$run$1.apply(FileAppender.scala:39)

at

org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:1311)

at

org.apache.spark.util.logging.FileAppender$$anon$1.run(FileAppender.scala:38)

===



The command df showed the following information at the worker node:



Filesystem   1K-blocks  Used Available Use% Mounted on

/dev/xvda1 8256920   8256456 0 100% /

tmpfs  7752012 0   7752012   0% /dev/shm

/dev/xvdb 30963708   1729652  27661192   6% /mnt



Does anybody know how to fix this?  Thanks.





Ey-Chih Chow







--

View this message in context: 
http://apache-spark-user-list.1001560.n3.nabble.com/no-space-left-at-worker-node-tp21545.html

Sent from the Apache Spark User List mailing list archive at Nabble.com.



-

To unsubscribe, e-mail: user-unsubscr...@spark.apache.org

For additional commands, e-mail: user-h...@spark.apache.org




  

  

Re: no space left at worker node

2015-02-08 Thread gen tang
Hi,

In fact, /dev/sdb is /dev/xvdb. It seems that there is no problem about
double mount. However, there is no information about /mnt2. You should
check whether /dev/sdc is well mounted or not.
The reply of Micheal is good solution about this type of problem. You can
check his site.

Cheers
Gen


On Sun, Feb 8, 2015 at 5:53 PM, ey-chih chow  wrote:

> Gen,
>
> Thanks for your information.  The content of /etc/fstab at the worker node
> (r3.large) is:
>
> #
> LABEL=/ /   ext4defaults,noatime  1   1
> tmpfs   /dev/shmtmpfs   defaults0   0
> devpts  /dev/ptsdevpts  gid=5,mode=620  0   0
> sysfs   /syssysfs   defaults0   0
> proc/proc   procdefaults0   0
> /dev/sdb/mntauto
>  defaults,noatime,nodiratime,comment=cloudconfig 0   0
> /dev/sdc/mnt2   auto
>  defaults,noatime,nodiratime,comment=cloudconfig 0   0
>
> There is no entry of /dev/xvdb.
>
>  Ey-Chih Chow
>
> ----------
> Date: Sun, 8 Feb 2015 12:09:37 +0100
> Subject: Re: no space left at worker node
> From: gen.tan...@gmail.com
> To: eyc...@hotmail.com
> CC: user@spark.apache.org
>
>
> Hi,
>
> I fact, I met this problem before. it is a bug of AWS. Which type of
> machine do you use?
>
> If I guess well, you can check the file /etc/fstab. There would be a
> double mount of /dev/xvdb.
> If yes, you should
> 1. stop hdfs
> 2. umount /dev/xvdb at /
> 3. restart hdfs
>
> Hope this could be helpful.
> Cheers
> Gen
>
>
>
> On Sun, Feb 8, 2015 at 8:16 AM, ey-chih chow  wrote:
>
> Hi,
>
> I submitted a spark job to an ec2 cluster, using spark-submit.  At a worker
> node, there is an exception of 'no space left on device' as follows.
>
> ==
> 15/02/08 01:53:38 ERROR logging.FileAppender: Error writing stream to file
> /root/spark/work/app-20150208014557-0003/0/stdout
> java.io.IOException: No space left on device
> at java.io.FileOutputStream.writeBytes(Native Method)
> at java.io.FileOutputStream.write(FileOutputStream.java:345)
> at
>
> org.apache.spark.util.logging.FileAppender.appendToFile(FileAppender.scala:92)
> at
>
> org.apache.spark.util.logging.FileAppender.appendStreamToFile(FileAppender.scala:72)
> at
>
> org.apache.spark.util.logging.FileAppender$$anon$1$$anonfun$run$1.apply$mcV$sp(FileAppender.scala:39)
> at
>
> org.apache.spark.util.logging.FileAppender$$anon$1$$anonfun$run$1.apply(FileAppender.scala:39)
> at
>
> org.apache.spark.util.logging.FileAppender$$anon$1$$anonfun$run$1.apply(FileAppender.scala:39)
> at
> org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:1311)
> at
>
> org.apache.spark.util.logging.FileAppender$$anon$1.run(FileAppender.scala:38)
> ===
>
> The command df showed the following information at the worker node:
>
> Filesystem   1K-blocks  Used Available Use% Mounted on
> /dev/xvda1 8256920   8256456 0 100% /
> tmpfs  7752012 0   7752012   0% /dev/shm
> /dev/xvdb 30963708   1729652  27661192   6% /mnt
>
> Does anybody know how to fix this?  Thanks.
>
>
> Ey-Chih Chow
>
>
>
> --
> View this message in context:
> http://apache-spark-user-list.1001560.n3.nabble.com/no-space-left-at-worker-node-tp21545.html
> Sent from the Apache Spark User List mailing list archive at Nabble.com.
>
> -
> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
> For additional commands, e-mail: user-h...@spark.apache.org
>
>
>


RE: no space left at worker node

2015-02-08 Thread ey-chih chow
Gen,
Thanks for your information.  The content of /etc/fstab at the worker node 
(r3.large) is:
#LABEL=/ /   ext4defaults,noatime  1   1tmpfs   /dev/shm
tmpfs   defaults0   0devpts  /dev/ptsdevpts  gid=5,mode=620  0  
 0sysfs   /syssysfs   defaults0   0proc/proc   
procdefaults0   0/dev/sdb/mntauto
defaults,noatime,nodiratime,comment=cloudconfig 0   0/dev/sdc/mnt2  
 autodefaults,noatime,nodiratime,comment=cloudconfig 0   0
There is no entry of /dev/xvdb.
 Ey-Chih Chow
Date: Sun, 8 Feb 2015 12:09:37 +0100
Subject: Re: no space left at worker node
From: gen.tan...@gmail.com
To: eyc...@hotmail.com
CC: user@spark.apache.org

Hi,
I fact, I met this problem before. it is a bug of AWS. Which type of machine do 
you use?
If I guess well, you can check the file /etc/fstab. There would be a double 
mount of /dev/xvdb.If yes, you should1. stop hdfs2. umount /dev/xvdb at / 3. 
restart hdfs
Hope this could be helpful.CheersGen


On Sun, Feb 8, 2015 at 8:16 AM, ey-chih chow  wrote:
Hi,



I submitted a spark job to an ec2 cluster, using spark-submit.  At a worker

node, there is an exception of 'no space left on device' as follows.



==

15/02/08 01:53:38 ERROR logging.FileAppender: Error writing stream to file

/root/spark/work/app-20150208014557-0003/0/stdout

java.io.IOException: No space left on device

at java.io.FileOutputStream.writeBytes(Native Method)

at java.io.FileOutputStream.write(FileOutputStream.java:345)

at

org.apache.spark.util.logging.FileAppender.appendToFile(FileAppender.scala:92)

at

org.apache.spark.util.logging.FileAppender.appendStreamToFile(FileAppender.scala:72)

at

org.apache.spark.util.logging.FileAppender$$anon$1$$anonfun$run$1.apply$mcV$sp(FileAppender.scala:39)

at

org.apache.spark.util.logging.FileAppender$$anon$1$$anonfun$run$1.apply(FileAppender.scala:39)

at

org.apache.spark.util.logging.FileAppender$$anon$1$$anonfun$run$1.apply(FileAppender.scala:39)

at

org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:1311)

at

org.apache.spark.util.logging.FileAppender$$anon$1.run(FileAppender.scala:38)

===



The command df showed the following information at the worker node:



Filesystem   1K-blocks  Used Available Use% Mounted on

/dev/xvda1 8256920   8256456 0 100% /

tmpfs  7752012 0   7752012   0% /dev/shm

/dev/xvdb 30963708   1729652  27661192   6% /mnt



Does anybody know how to fix this?  Thanks.





Ey-Chih Chow







--

View this message in context: 
http://apache-spark-user-list.1001560.n3.nabble.com/no-space-left-at-worker-node-tp21545.html

Sent from the Apache Spark User List mailing list archive at Nabble.com.



-

To unsubscribe, e-mail: user-unsubscr...@spark.apache.org

For additional commands, e-mail: user-h...@spark.apache.org




  

Re: no space left at worker node

2015-02-08 Thread gen tang
Hi,

I fact, I met this problem before. it is a bug of AWS. Which type of
machine do you use?

If I guess well, you can check the file /etc/fstab. There would be a double
mount of /dev/xvdb.
If yes, you should
1. stop hdfs
2. umount /dev/xvdb at /
3. restart hdfs

Hope this could be helpful.
Cheers
Gen



On Sun, Feb 8, 2015 at 8:16 AM, ey-chih chow  wrote:

> Hi,
>
> I submitted a spark job to an ec2 cluster, using spark-submit.  At a worker
> node, there is an exception of 'no space left on device' as follows.
>
> ==
> 15/02/08 01:53:38 ERROR logging.FileAppender: Error writing stream to file
> /root/spark/work/app-20150208014557-0003/0/stdout
> java.io.IOException: No space left on device
> at java.io.FileOutputStream.writeBytes(Native Method)
> at java.io.FileOutputStream.write(FileOutputStream.java:345)
> at
>
> org.apache.spark.util.logging.FileAppender.appendToFile(FileAppender.scala:92)
> at
>
> org.apache.spark.util.logging.FileAppender.appendStreamToFile(FileAppender.scala:72)
> at
>
> org.apache.spark.util.logging.FileAppender$$anon$1$$anonfun$run$1.apply$mcV$sp(FileAppender.scala:39)
> at
>
> org.apache.spark.util.logging.FileAppender$$anon$1$$anonfun$run$1.apply(FileAppender.scala:39)
> at
>
> org.apache.spark.util.logging.FileAppender$$anon$1$$anonfun$run$1.apply(FileAppender.scala:39)
> at
> org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:1311)
> at
>
> org.apache.spark.util.logging.FileAppender$$anon$1.run(FileAppender.scala:38)
> ===
>
> The command df showed the following information at the worker node:
>
> Filesystem   1K-blocks  Used Available Use% Mounted on
> /dev/xvda1 8256920   8256456 0 100% /
> tmpfs  7752012 0   7752012   0% /dev/shm
> /dev/xvdb 30963708   1729652  27661192   6% /mnt
>
> Does anybody know how to fix this?  Thanks.
>
>
> Ey-Chih Chow
>
>
>
> --
> View this message in context:
> http://apache-spark-user-list.1001560.n3.nabble.com/no-space-left-at-worker-node-tp21545.html
> Sent from the Apache Spark User List mailing list archive at Nabble.com.
>
> -
> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
> For additional commands, e-mail: user-h...@spark.apache.org
>
>