Looks like this is caused by issue SPARK-5008:
https://issues.apache.org/jira/browse/SPARK-5008

On 13 February 2015 at 19:04, Joe Wass <jw...@crossref.org> wrote:

> I've updated to Spark 1.2.0 and the EC2 and the persistent-hdfs behaviour
> appears to have changed.
>
> My launch script is
>
> spark-1.2.0-bin-hadoop2.4/ec2/spark-ec2 --instance-type=m3.xlarge -s 5
> --ebs-vol-size=1000 launch myproject
>
> When I ssh into master I get:
>
> $ df -h
> Filesystem            Size  Used Avail Use% Mounted on
> /dev/xvda1            7.9G  2.9G  5.0G  37% /
> tmpfs                 7.3G     0  7.3G   0% /dev/shm
> /dev/xvdb              37G  1.3G   34G   4% /mnt
> /dev/xvdc              37G  177M   35G   1% /mnt2
> /dev/xvds            1000G   33M 1000G   1% /vol0
>
> that /vol0 is the place I want (and assume) persistent-hdfs to go. But
> when I look at the size I get:
>
> $ persistent-hdfs/bin/start-all.sh
> $ persistent-hdfs/bin/hadoop dfsadmin -report
> Warning: $HADOOP_HOME is deprecated.
>
> Configured Capacity: 42275430400 (39.37 GB)
> Present Capacity: 26448744448 (24.63 GB)
> DFS Remaining: 26448601088 (24.63 GB)
> DFS Used: 143360 (140 KB)
> DFS Used%: 0%
> Under replicated blocks: 0
> Blocks with corrupt replicas: 0
> Missing blocks: 0
>
> -------------------------------------------------
> Datanodes available: 5 (5 total, 0 dead)
>
> Name: 10.46.11.156:60010
> Decommission Status : Normal
> Configured Capacity: 8455086080 (7.87 GB)
> DFS Used: 28672 (28 KB)
> Non DFS Used: 3165372416 (2.95 GB)
> DFS Remaining: 5289684992(4.93 GB)
> DFS Used%: 0%
> DFS Remaining%: 62.56%
> Last contact: Fri Feb 13 17:41:46 UTC 2015
>
>
> Name: 10.41.51.155:60010
> Decommission Status : Normal
> Configured Capacity: 8455086080 (7.87 GB)
> DFS Used: 28672 (28 KB)
> Non DFS Used: 3165364224 (2.95 GB)
> DFS Remaining: 5289693184(4.93 GB)
> DFS Used%: 0%
> DFS Remaining%: 62.56%
> Last contact: Fri Feb 13 17:41:46 UTC 2015
>
>
> Name: 10.38.30.254:60010
> Decommission Status : Normal
> Configured Capacity: 8455086080 (7.87 GB)
> DFS Used: 28672 (28 KB)
> Non DFS Used: 3165249536 (2.95 GB)
> DFS Remaining: 5289807872(4.93 GB)
> DFS Used%: 0%
> DFS Remaining%: 62.56%
> Last contact: Fri Feb 13 17:41:46 UTC 2015
>
>
> Name: 10.204.134.84:60010
> Decommission Status : Normal
> Configured Capacity: 8455086080 (7.87 GB)
> DFS Used: 28672 (28 KB)
> Non DFS Used: 3165343744 (2.95 GB)
> DFS Remaining: 5289713664(4.93 GB)
> DFS Used%: 0%
> DFS Remaining%: 62.56%
> Last contact: Fri Feb 13 17:41:46 UTC 2015
>
>
> Name: 10.33.15.134:60010
> Decommission Status : Normal
> Configured Capacity: 8455086080 (7.87 GB)
> DFS Used: 28672 (28 KB)
> Non DFS Used: 3165356032 (2.95 GB)
> DFS Remaining: 5289701376(4.93 GB)
> DFS Used%: 0%
> DFS Remaining%: 62.56%
> Last contact: Fri Feb 13 17:41:46 UTC 2015
>
>
> That's tiny. My suspicions are aroused when I see:
>
> $ ls /vol
> persistent-hdfs
>
> /vol is on the small /dev/xvda1 not the large EBS /dev/xvds
>
> I thought I'd be able to edit persistent-hdfs/conf/core-site.xml to change
> the volume:
>
> <property>
>   <name>hadoop.tmp.dir</name>
>   <value>/vol0/persistent-hdfs</value>  <!-- was /vol/persistent-hdfs -->
> </property>
>
> And then
>
> persistent-hdfs/bin/stop-all.sh && persistent-hdfs/bin/start-all.sh
>
> but when I do that, the persistent HDFS won't start for whatever reason. I
> can't run
>
> $ persistent-hdfs/bin/hadoop dfsadmin -report
>
> 15/02/13 18:50:25 INFO ipc.Client: Retrying connect to server:
> ec2-54-70-252-81.us-west-2.compute.amazonaws.com/10.23.161.84:9010.
> Already tried 0 time(s).
> 15/02/13 18:50:26 INFO ipc.Client: Retrying connect to server:
> ec2-54-70-252-81.us-west-2.compute.amazonaws.com/10.23.161.84:9010.
> Already tried 1 time(s).
> 15/02/13 18:50:27 INFO ipc.Client: Retrying connect to server:
> ec2-54-70-252-81.us-west-2.compute.amazonaws.com/10.23.161.84:9010.
> Already tried 2 time(s).
>
> So, it looks like I can't use the EBS for persistent-hdfs. I was doing it
> before, so something must have changed in the last couple of weeks (last
> time I was using 1.1.0).
>
> Is this a bug? Has the behaviour of AWS changed? Am I doing something
> stupid? How do I fix it?
>
> Thanks in advance!
>
> Joe
>
>
>

Reply via email to