[
https://issues.apache.org/jira/browse/WHIRR-96?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12913076#action_12913076
]
Zach Bailey edited comment on WHIRR-96 at 9/21/10 12:57 PM:
------------------------------------------------------------
Followed the instructions on the page above to the letter besides using a
different cluster name. When starting the cluster, the instances start
correctly and the volumes are attached fine:
{quote}
Waiting for 1 instances in role nn,snn,jt to start
.....................................nn,snn,jt instances started
Waiting for 2 instances in role dn,tt to start
.......................................dn,tt instances started
Waiting 10 seconds before attaching storage
Attaching storage to i-eb5a6581
Attaching vol-9a3e17f3 to i-eb5a6581
Attaching vol-9c3e17f5 to i-eb5a6581
Attaching storage to i-df5a65b5
Attaching vol-923e17fb to i-df5a65b5
Attaching vol-943e17fd to i-df5a65b5
Attaching storage to i-dd5a65b7
Attaching vol-963e17ff to i-dd5a65b7
Attaching vol-68311801 to i-dd5a65b7
nn vol-9a3e17f3 250 snap-76cafa1d us-east-1a in-use
2010-09-21T14:24:20.000Z None
nn vol-9c3e17f5 250 snap-76cafa1d us-east-1a in-use
2010-09-21T14:24:20.000Z None
dn vol-923e17fb 250 snap-76cafa1d us-east-1a in-use
2010-09-21T14:24:39.000Z None
dn vol-943e17fd 250 snap-76cafa1d us-east-1a in-use
2010-09-21T14:24:42.000Z None
dn vol-963e17ff 250 snap-76cafa1d us-east-1a in-use
2010-09-21T14:24:45.000Z None
dn vol-68311801 250 snap-76cafa1d us-east-1a in-use
2010-09-21T14:24:46.000Z None
{quote}
However, the next part where the scripts wait for the jobtracker to start time
out:
Waiting for jobtracker to start
.....................................................................................................................................................................................................................................................................................................Timeout
while waiting for Hadoop to start. Please check logs on cluster.
Checking /var/log/messages on the cluster NN yield the following error spamming
the logs approx every 10 seconds:
Sep 21 16:35:59 domU-12-31-39-06-18-D1 ec2:
#############################################################
Sep 21 16:37:21 domU-12-31-39-06-18-D1 kernel: [ 88.139988] GFS2: path_lookup
on /dev/sdk returned error -2
Sep 21 16:37:31 domU-12-31-39-06-18-D1 kernel: [ 98.145469] GFS2: path_lookup
on /dev/sdk returned error -2
Sep 21 16:37:41 domU-12-31-39-06-18-D1 kernel: [ 108.151421] GFS2: path_lookup
on /dev/sdk returned error -2
Sep 21 16:37:51 domU-12-31-39-06-18-D1 kernel: [ 118.178490] GFS2: gfs2 mount
does not exist
Sep 21 16:38:02 domU-12-31-39-06-18-D1 kernel: [ 128.205265] GFS2: gfs2 mount
does not exist
Sep 21 16:38:12 domU-12-31-39-06-18-D1 kernel: [ 138.228104] GFS2: gfs2 mount
does not exist
Sep 21 16:38:22 domU-12-31-39-06-18-D1 kernel: [ 148.252729] GFS2: gfs2 mount
does not exist
Sep 21 16:38:32 domU-12-31-39-06-18-D1 kernel: [ 158.277586] GFS2: gfs2 mount
does not exist
Sep 21 16:38:42 domU-12-31-39-06-18-D1 kernel: [ 168.311393] GFS2: gfs2 mount
does not exist
Sep 21 16:38:52 domU-12-31-39-06-18-D1 kernel: [ 178.329587] GFS2: gfs2 mount
does not exist
Sep 21 16:39:02 domU-12-31-39-06-18-D1 kernel: [ 188.349149] GFS2: gfs2 mount
does not exist
Sep 21 16:39:12 domU-12-31-39-06-18-D1 kernel: [ 198.369786] GFS2: gfs2 mount
does not exist
Sep 21 16:39:22 domU-12-31-39-06-18-D1 kernel: [ 208.389564] GFS2: gfs2 mount
does not exist
Sep 21 16:39:32 domU-12-31-39-06-18-D1 kernel: [ 218.409901] GFS2: gfs2 mount
does not exist
was (Author: znbailey):
Followed the instructions on the page above to the letter besides using a
different cluster name. When starting the cluster, the instances start
correctly and the volumes are attached fine:
Waiting for 1 instances in role nn,snn,jt to start
.....................................nn,snn,jt instances started
Waiting for 2 instances in role dn,tt to start
.......................................dn,tt instances started
Waiting 10 seconds before attaching storage
Attaching storage to i-eb5a6581
Attaching vol-9a3e17f3 to i-eb5a6581
Attaching vol-9c3e17f5 to i-eb5a6581
Attaching storage to i-df5a65b5
Attaching vol-923e17fb to i-df5a65b5
Attaching vol-943e17fd to i-df5a65b5
Attaching storage to i-dd5a65b7
Attaching vol-963e17ff to i-dd5a65b7
Attaching vol-68311801 to i-dd5a65b7
nn vol-9a3e17f3 250 snap-76cafa1d us-east-1a in-use
2010-09-21T14:24:20.000Z None
nn vol-9c3e17f5 250 snap-76cafa1d us-east-1a in-use
2010-09-21T14:24:20.000Z None
dn vol-923e17fb 250 snap-76cafa1d us-east-1a in-use
2010-09-21T14:24:39.000Z None
dn vol-943e17fd 250 snap-76cafa1d us-east-1a in-use
2010-09-21T14:24:42.000Z None
dn vol-963e17ff 250 snap-76cafa1d us-east-1a in-use
2010-09-21T14:24:45.000Z None
dn vol-68311801 250 snap-76cafa1d us-east-1a in-use
2010-09-21T14:24:46.000Z None
However, the next part where the scripts wait for the jobtracker to start time
out:
Waiting for jobtracker to start
.....................................................................................................................................................................................................................................................................................................Timeout
while waiting for Hadoop to start. Please check logs on cluster.}}
Checking /var/log/messages on the cluster NN yield the following error spamming
the logs approx every 10 seconds:
Sep 21 16:35:59 domU-12-31-39-06-18-D1 ec2:
#############################################################
Sep 21 16:37:21 domU-12-31-39-06-18-D1 kernel: [ 88.139988] GFS2: path_lookup
on /dev/sdk returned error -2
Sep 21 16:37:31 domU-12-31-39-06-18-D1 kernel: [ 98.145469] GFS2: path_lookup
on /dev/sdk returned error -2
Sep 21 16:37:41 domU-12-31-39-06-18-D1 kernel: [ 108.151421] GFS2: path_lookup
on /dev/sdk returned error -2
Sep 21 16:37:51 domU-12-31-39-06-18-D1 kernel: [ 118.178490] GFS2: gfs2 mount
does not exist
Sep 21 16:38:02 domU-12-31-39-06-18-D1 kernel: [ 128.205265] GFS2: gfs2 mount
does not exist
Sep 21 16:38:12 domU-12-31-39-06-18-D1 kernel: [ 138.228104] GFS2: gfs2 mount
does not exist
Sep 21 16:38:22 domU-12-31-39-06-18-D1 kernel: [ 148.252729] GFS2: gfs2 mount
does not exist
Sep 21 16:38:32 domU-12-31-39-06-18-D1 kernel: [ 158.277586] GFS2: gfs2 mount
does not exist
Sep 21 16:38:42 domU-12-31-39-06-18-D1 kernel: [ 168.311393] GFS2: gfs2 mount
does not exist
Sep 21 16:38:52 domU-12-31-39-06-18-D1 kernel: [ 178.329587] GFS2: gfs2 mount
does not exist
Sep 21 16:39:02 domU-12-31-39-06-18-D1 kernel: [ 188.349149] GFS2: gfs2 mount
does not exist
Sep 21 16:39:12 domU-12-31-39-06-18-D1 kernel: [ 198.369786] GFS2: gfs2 mount
does not exist
Sep 21 16:39:22 domU-12-31-39-06-18-D1 kernel: [ 208.389564] GFS2: gfs2 mount
does not exist
Sep 21 16:39:32 domU-12-31-39-06-18-D1 kernel: [ 218.409901] GFS2: gfs2 mount
does not exist
> EBS cluster doesn't work with CDH3 on Lucid
> -------------------------------------------
>
> Key: WHIRR-96
> URL: https://issues.apache.org/jira/browse/WHIRR-96
> Project: Whirr
> Issue Type: Bug
> Components: contrib/python
> Reporter: Tom White
>
> Following the instructions with
> https://docs.cloudera.com/display/DOC/Using+Persistent+Clusters and
> ami-2d4aa444 gives "Timeout while waiting for Hadoop to start. Please check
> logs on cluster."
--
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.