Re: java.io.IOException: Function not implemented

2010-03-31 Thread Steve Loughran

Edson Ramiro wrote:

May be it's a bug.

I'm not the admin. : (

so, I'll talk to him and may be he install a 2.6.32.9 in another node to
test  : )

Thanks

Edson Ramiro


On 30 March 2010 20:00, Todd Lipcon t...@cloudera.com wrote:


Hi Edson,

I noticed that only the h01 nodes are running 2.6.32.9, the other broken
DNs
are 2.6.32.10.

Is there some reason you are running a kernel that is literally 2 weeks
old?
I wouldn't be at all surprised if there were a bug here, or some issue with
your Debian unstable distribution...



If you are running the SCM trunk of the OS, you are part of the dev 
team. They will be grateful for the bugs you find and fix, but you get 
to find and fix them. In Ant one bugrep was touch stopped setting 
dates in the past, turned out that on the debian nightly builds, you 
couldn't touch any file into the past...


-steve





Re: java.io.IOException: Function not implemented

2010-03-30 Thread Steve Loughran

Edson Ramiro wrote:

I'm not involved with Debian community :(


I think you are now...


Re: java.io.IOException: Function not implemented

2010-03-30 Thread Edson Ramiro
Hi all,

Thanks for help Todd and Steve,

I configured Hadoop (0.20.2) again and I'm getting the same error (Function
not implemented).

Do you think it's a Hadoop bug?

This is the situation:

I've 28 nodes where just four are running the datanode.

In all other nodes the tasktracker in running ok.

The NN and JT are running ok.

The configuration of the machines is the same, its a nfs shared home.

In all machines the Java version is 1.6.0_17.

This is the kernel version of the nodes, note that are two versions and in
both the
datanode doesn't work. Just in the h0* machines.

ram...@lcpad:~/hadoop-0.20.2$ ./bin/slaves.sh uname -a  | sort
a01: Linux a01 2.6.27.11 #4 Fri Jan 16 22:32:46 BRST 2009 x86_64 GNU/Linux
a02: Linux a02 2.6.27.11 #4 Fri Jan 16 22:32:46 BRST 2009 x86_64 GNU/Linux
a03: Linux a03 2.6.27.11 #4 Fri Jan 16 22:32:46 BRST 2009 x86_64 GNU/Linux
a04: Linux a04 2.6.27.11 #4 Fri Jan 16 22:32:46 BRST 2009 x86_64 GNU/Linux
a05: Linux a05 2.6.27.11 #4 Fri Jan 16 22:32:46 BRST 2009 x86_64 GNU/Linux
a06: Linux a06 2.6.27.11 #4 Fri Jan 16 22:32:46 BRST 2009 x86_64 GNU/Linux
a07: Linux a07 2.6.27.11 #4 Fri Jan 16 22:32:46 BRST 2009 x86_64 GNU/Linux
a09: Linux a09 2.6.27.11 #4 Fri Jan 16 22:32:46 BRST 2009 x86_64 GNU/Linux
a10: Linux a10 2.6.27.11 #4 Fri Jan 16 22:32:46 BRST 2009 x86_64 GNU/Linux
ag06: Linux ag06 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
GNU/Linux
ag07: Linux ag07 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
GNU/Linux
bl02: Linux bl02 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
GNU/Linux
bl03: Linux bl03 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
GNU/Linux
bl04: Linux bl04 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
GNU/Linux
bl06: Linux bl06 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
GNU/Linux
bl07: Linux bl07 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
GNU/Linux
ct02: Linux ct02 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
GNU/Linux
ct03: Linux ct03 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
GNU/Linux
ct04: Linux ct04 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
GNU/Linux
ct06: Linux ct06 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
GNU/Linux
h01: Linux h01 2.6.32.9 #2 SMP Sat Mar 6 19:09:13 BRT 2010 x86_64 GNU/Linux
h02: Linux h02 2.6.32.9 #2 SMP Sat Mar 6 19:09:13 BRT 2010 x86_64 GNU/Linux
h03: Linux h03 2.6.32.9 #2 SMP Sat Mar 6 19:09:13 BRT 2010 x86_64 GNU/Linux
h04: Linux h04 2.6.32.9 #2 SMP Sat Mar 6 19:09:13 BRT 2010 x86_64 GNU/Linux
sd02: Linux sd02 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
GNU/Linux
sd05: Linux sd05 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
GNU/Linux
sd06: Linux sd06 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
GNU/Linux
sd07: Linux sd07 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
GNU/Linux


These are the java processes running on each clients.
Jjust the h0* machines are running ok.

ram...@lcpad:~/hadoop-0.20.2$ ./bin/slaves.sh pgrep -lc java | sort
a01: 1
a02: 1
a03: 1
a04: 1
a05: 1
a06: 1
a07: 1
a09: 1
a10: 1
ag06: 1
ag07: 1
bl02: 1
bl03: 1
bl04: 1
bl06: 1
bl07: 1
ct02: 1
ct03: 1
ct04: 1
ct06: 1
h01: 2
h02: 2
h03: 2
h04: 2
sd02: 1
sd05: 1
sd06: 1
sd07: 1

This is my configuration:

ram...@lcpad:~/hadoop-0.20.2$ cat conf/*site*
?xml version=1.0?
?xml-stylesheet type=text/xsl href=configuration.xsl?

!-- Put site-specific property overrides in this file. --

configuration
property
namefs.default.name/name
valuehdfs://lcpad:9000/value
/property
/configuration
?xml version=1.0?
?xml-stylesheet type=text/xsl href=configuration.xsl?

!-- Put site-specific property overrides in this file. --

configuration
property
namedfs.replication/name
value1/value
/property
/configuration
?xml version=1.0?
?xml-stylesheet type=text/xsl href=configuration.xsl?

!-- Put site-specific property overrides in this file. --

configuration
  property
namemapred.job.tracker/name
valuelcpad:9001/value
  /property
/configuration

Thanks in Advance,

Edson Ramiro


On 30 March 2010 05:58, Steve Loughran ste...@apache.org wrote:

 Edson Ramiro wrote:

 I'm not involved with Debian community :(


 I think you are now...



Re: java.io.IOException: Function not implemented

2010-03-30 Thread Todd Lipcon
Hi Edson,

I noticed that only the h01 nodes are running 2.6.32.9, the other broken DNs
are 2.6.32.10.

Is there some reason you are running a kernel that is literally 2 weeks old?
I wouldn't be at all surprised if there were a bug here, or some issue with
your Debian unstable distribution...

-Todd

On Tue, Mar 30, 2010 at 3:54 PM, Edson Ramiro erlfi...@gmail.com wrote:

 Hi all,

 Thanks for help Todd and Steve,

 I configured Hadoop (0.20.2) again and I'm getting the same error (Function
 not implemented).

 Do you think it's a Hadoop bug?

 This is the situation:

 I've 28 nodes where just four are running the datanode.

 In all other nodes the tasktracker in running ok.

 The NN and JT are running ok.

 The configuration of the machines is the same, its a nfs shared home.

 In all machines the Java version is 1.6.0_17.

 This is the kernel version of the nodes, note that are two versions and in
 both the
 datanode doesn't work. Just in the h0* machines.

 ram...@lcpad:~/hadoop-0.20.2$ ./bin/slaves.sh uname -a  | sort
 a01: Linux a01 2.6.27.11 #4 Fri Jan 16 22:32:46 BRST 2009 x86_64 GNU/Linux
 a02: Linux a02 2.6.27.11 #4 Fri Jan 16 22:32:46 BRST 2009 x86_64 GNU/Linux
 a03: Linux a03 2.6.27.11 #4 Fri Jan 16 22:32:46 BRST 2009 x86_64 GNU/Linux
 a04: Linux a04 2.6.27.11 #4 Fri Jan 16 22:32:46 BRST 2009 x86_64 GNU/Linux
 a05: Linux a05 2.6.27.11 #4 Fri Jan 16 22:32:46 BRST 2009 x86_64 GNU/Linux
 a06: Linux a06 2.6.27.11 #4 Fri Jan 16 22:32:46 BRST 2009 x86_64 GNU/Linux
 a07: Linux a07 2.6.27.11 #4 Fri Jan 16 22:32:46 BRST 2009 x86_64 GNU/Linux
 a09: Linux a09 2.6.27.11 #4 Fri Jan 16 22:32:46 BRST 2009 x86_64 GNU/Linux
 a10: Linux a10 2.6.27.11 #4 Fri Jan 16 22:32:46 BRST 2009 x86_64 GNU/Linux
 ag06: Linux ag06 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
 GNU/Linux
 ag07: Linux ag07 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
 GNU/Linux
 bl02: Linux bl02 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
 GNU/Linux
 bl03: Linux bl03 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
 GNU/Linux
 bl04: Linux bl04 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
 GNU/Linux
 bl06: Linux bl06 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
 GNU/Linux
 bl07: Linux bl07 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
 GNU/Linux
 ct02: Linux ct02 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
 GNU/Linux
 ct03: Linux ct03 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
 GNU/Linux
 ct04: Linux ct04 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
 GNU/Linux
 ct06: Linux ct06 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
 GNU/Linux
 h01: Linux h01 2.6.32.9 #2 SMP Sat Mar 6 19:09:13 BRT 2010 x86_64 GNU/Linux
 h02: Linux h02 2.6.32.9 #2 SMP Sat Mar 6 19:09:13 BRT 2010 x86_64 GNU/Linux
 h03: Linux h03 2.6.32.9 #2 SMP Sat Mar 6 19:09:13 BRT 2010 x86_64 GNU/Linux
 h04: Linux h04 2.6.32.9 #2 SMP Sat Mar 6 19:09:13 BRT 2010 x86_64 GNU/Linux
 sd02: Linux sd02 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
 GNU/Linux
 sd05: Linux sd05 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
 GNU/Linux
 sd06: Linux sd06 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
 GNU/Linux
 sd07: Linux sd07 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
 GNU/Linux


 These are the java processes running on each clients.
 Jjust the h0* machines are running ok.

 ram...@lcpad:~/hadoop-0.20.2$ ./bin/slaves.sh pgrep -lc java | sort
 a01: 1
 a02: 1
 a03: 1
 a04: 1
 a05: 1
 a06: 1
 a07: 1
 a09: 1
 a10: 1
 ag06: 1
 ag07: 1
 bl02: 1
 bl03: 1
 bl04: 1
 bl06: 1
 bl07: 1
 ct02: 1
 ct03: 1
 ct04: 1
 ct06: 1
 h01: 2
 h02: 2
 h03: 2
 h04: 2
 sd02: 1
 sd05: 1
 sd06: 1
 sd07: 1

 This is my configuration:

 ram...@lcpad:~/hadoop-0.20.2$ cat conf/*site*
 ?xml version=1.0?
 ?xml-stylesheet type=text/xsl href=configuration.xsl?

 !-- Put site-specific property overrides in this file. --

 configuration
 property
 namefs.default.name/name
 valuehdfs://lcpad:9000/value
 /property
 /configuration
 ?xml version=1.0?
 ?xml-stylesheet type=text/xsl href=configuration.xsl?

 !-- Put site-specific property overrides in this file. --

 configuration
 property
 namedfs.replication/name
 value1/value
 /property
 /configuration
 ?xml version=1.0?
 ?xml-stylesheet type=text/xsl href=configuration.xsl?

 !-- Put site-specific property overrides in this file. --

 configuration
  property
namemapred.job.tracker/name
valuelcpad:9001/value
  /property
 /configuration

 Thanks in Advance,

 Edson Ramiro


 On 30 March 2010 05:58, Steve Loughran ste...@apache.org wrote:

  Edson Ramiro wrote:
 
  I'm not involved with Debian community :(
 
 
  I think you are now...
 




-- 
Todd Lipcon
Software Engineer, Cloudera


Re: java.io.IOException: Function not implemented

2010-03-30 Thread Edson Ramiro
May be it's a bug.

I'm not the admin. : (

so, I'll talk to him and may be he install a 2.6.32.9 in another node to
test  : )

Thanks

Edson Ramiro


On 30 March 2010 20:00, Todd Lipcon t...@cloudera.com wrote:

 Hi Edson,

 I noticed that only the h01 nodes are running 2.6.32.9, the other broken
 DNs
 are 2.6.32.10.

 Is there some reason you are running a kernel that is literally 2 weeks
 old?
 I wouldn't be at all surprised if there were a bug here, or some issue with
 your Debian unstable distribution...

 -Todd

 On Tue, Mar 30, 2010 at 3:54 PM, Edson Ramiro erlfi...@gmail.com wrote:

  Hi all,
 
  Thanks for help Todd and Steve,
 
  I configured Hadoop (0.20.2) again and I'm getting the same error
 (Function
  not implemented).
 
  Do you think it's a Hadoop bug?
 
  This is the situation:
 
  I've 28 nodes where just four are running the datanode.
 
  In all other nodes the tasktracker in running ok.
 
  The NN and JT are running ok.
 
  The configuration of the machines is the same, its a nfs shared home.
 
  In all machines the Java version is 1.6.0_17.
 
  This is the kernel version of the nodes, note that are two versions and
 in
  both the
  datanode doesn't work. Just in the h0* machines.
 
  ram...@lcpad:~/hadoop-0.20.2$ ./bin/slaves.sh uname -a  | sort
  a01: Linux a01 2.6.27.11 #4 Fri Jan 16 22:32:46 BRST 2009 x86_64
 GNU/Linux
  a02: Linux a02 2.6.27.11 #4 Fri Jan 16 22:32:46 BRST 2009 x86_64
 GNU/Linux
  a03: Linux a03 2.6.27.11 #4 Fri Jan 16 22:32:46 BRST 2009 x86_64
 GNU/Linux
  a04: Linux a04 2.6.27.11 #4 Fri Jan 16 22:32:46 BRST 2009 x86_64
 GNU/Linux
  a05: Linux a05 2.6.27.11 #4 Fri Jan 16 22:32:46 BRST 2009 x86_64
 GNU/Linux
  a06: Linux a06 2.6.27.11 #4 Fri Jan 16 22:32:46 BRST 2009 x86_64
 GNU/Linux
  a07: Linux a07 2.6.27.11 #4 Fri Jan 16 22:32:46 BRST 2009 x86_64
 GNU/Linux
  a09: Linux a09 2.6.27.11 #4 Fri Jan 16 22:32:46 BRST 2009 x86_64
 GNU/Linux
  a10: Linux a10 2.6.27.11 #4 Fri Jan 16 22:32:46 BRST 2009 x86_64
 GNU/Linux
  ag06: Linux ag06 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
  GNU/Linux
  ag07: Linux ag07 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
  GNU/Linux
  bl02: Linux bl02 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
  GNU/Linux
  bl03: Linux bl03 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
  GNU/Linux
  bl04: Linux bl04 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
  GNU/Linux
  bl06: Linux bl06 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
  GNU/Linux
  bl07: Linux bl07 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
  GNU/Linux
  ct02: Linux ct02 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
  GNU/Linux
  ct03: Linux ct03 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
  GNU/Linux
  ct04: Linux ct04 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
  GNU/Linux
  ct06: Linux ct06 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
  GNU/Linux
  h01: Linux h01 2.6.32.9 #2 SMP Sat Mar 6 19:09:13 BRT 2010 x86_64
 GNU/Linux
  h02: Linux h02 2.6.32.9 #2 SMP Sat Mar 6 19:09:13 BRT 2010 x86_64
 GNU/Linux
  h03: Linux h03 2.6.32.9 #2 SMP Sat Mar 6 19:09:13 BRT 2010 x86_64
 GNU/Linux
  h04: Linux h04 2.6.32.9 #2 SMP Sat Mar 6 19:09:13 BRT 2010 x86_64
 GNU/Linux
  sd02: Linux sd02 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
  GNU/Linux
  sd05: Linux sd05 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
  GNU/Linux
  sd06: Linux sd06 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
  GNU/Linux
  sd07: Linux sd07 2.6.32.10 #1 SMP Tue Mar 16 10:17:30 BRT 2010 x86_64
  GNU/Linux
 
 
  These are the java processes running on each clients.
  Jjust the h0* machines are running ok.
 
  ram...@lcpad:~/hadoop-0.20.2$ ./bin/slaves.sh pgrep -lc java | sort
  a01: 1
  a02: 1
  a03: 1
  a04: 1
  a05: 1
  a06: 1
  a07: 1
  a09: 1
  a10: 1
  ag06: 1
  ag07: 1
  bl02: 1
  bl03: 1
  bl04: 1
  bl06: 1
  bl07: 1
  ct02: 1
  ct03: 1
  ct04: 1
  ct06: 1
  h01: 2
  h02: 2
  h03: 2
  h04: 2
  sd02: 1
  sd05: 1
  sd06: 1
  sd07: 1
 
  This is my configuration:
 
  ram...@lcpad:~/hadoop-0.20.2$ cat conf/*site*
  ?xml version=1.0?
  ?xml-stylesheet type=text/xsl href=configuration.xsl?
 
  !-- Put site-specific property overrides in this file. --
 
  configuration
  property
  namefs.default.name/name
  valuehdfs://lcpad:9000/value
  /property
  /configuration
  ?xml version=1.0?
  ?xml-stylesheet type=text/xsl href=configuration.xsl?
 
  !-- Put site-specific property overrides in this file. --
 
  configuration
  property
  namedfs.replication/name
  value1/value
  /property
  /configuration
  ?xml version=1.0?
  ?xml-stylesheet type=text/xsl href=configuration.xsl?
 
  !-- Put site-specific property overrides in this file. --
 
  configuration
   property
 namemapred.job.tracker/name
 valuelcpad:9001/value
   /property
  /configuration
 
  Thanks in Advance,
 
  Edson Ramiro
 
 
  On 30 March 2010 05:58, Steve Loughran ste...@apache.org wrote:
 
   Edson Ramiro wrote:
  
   I'm not involved 

java.io.IOException: Function not implemented

2010-03-29 Thread Edson Ramiro
Hi all,

I'm trying to install Hadoop on a cluster, but I'm getting this error.

I'm using java version 1.6.0_17 and hadoop-0.20.1+169.56.tar.gz from
Cloudera.

Its running in a NFS home shared between the nodes and masters.

The NameNode works well, but all nodes try to connect and fail.

Any Idea ?

Thanks in Advance.

== logs/hadoop-ramiro-datanode-a05.log ==
2010-03-29 15:56:00,168 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: lcpad/192.168.1.51:9000. Already tried 0 time(s).
2010-03-29 15:56:01,172 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: lcpad/192.168.1.51:9000. Already tried 1 time(s).
2010-03-29 15:56:02,176 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: lcpad/192.168.1.51:9000. Already tried 2 time(s).
2010-03-29 15:56:03,180 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: lcpad/192.168.1.51:9000. Already tried 3 time(s).
2010-03-29 15:56:04,184 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: lcpad/192.168.1.51:9000. Already tried 4 time(s).
2010-03-29 15:56:05,188 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: lcpad/192.168.1.51:9000. Already tried 5 time(s).
2010-03-29 15:56:06,192 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: lcpad/192.168.1.51:9000. Already tried 6 time(s).
2010-03-29 15:56:07,196 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: lcpad/192.168.1.51:9000. Already tried 7 time(s).
2010-03-29 15:56:08,200 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: lcpad/192.168.1.51:9000. Already tried 8 time(s).
2010-03-29 15:56:09,204 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: lcpad/192.168.1.51:9000. Already tried 9 time(s).
2010-03-29 15:56:09,204 ERROR
org.apache.hadoop.hdfs.server.datanode.DataNode: java.io.IOException: Call
to lcpad/192.168.1.51:9000 failed on local exception: java.io.IOException:
Function not implemented
at org.apache.hadoop.ipc.Client.wrapException(Client.java:775)
at org.apache.hadoop.ipc.Client.call(Client.java:743)
at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:220)
at $Proxy4.getProtocolVersion(Unknown Source)
at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:359)
at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:346)
at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:383)
at org.apache.hadoop.ipc.RPC.waitForProxy(RPC.java:314)
at org.apache.hadoop.ipc.RPC.waitForProxy(RPC.java:291)
at
org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:278)
at
org.apache.hadoop.hdfs.server.datanode.DataNode.init(DataNode.java:225)
at
org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:1309)
at
org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:1264)
at
org.apache.hadoop.hdfs.server.datanode.DataNode.createDataNode(DataNode.java:1272)
at
org.apache.hadoop.hdfs.server.datanode.DataNode.main(DataNode.java:1394)
Caused by: java.io.IOException: Function not implemented
at sun.nio.ch.EPollArrayWrapper.epollCreate(Native Method)
at sun.nio.ch.EPollArrayWrapper.init(EPollArrayWrapper.java:68)
at sun.nio.ch.EPollSelectorImpl.init(EPollSelectorImpl.java:52)
at
sun.nio.ch.EPollSelectorProvider.openSelector(EPollSelectorProvider.java:18)
at
org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.get(SocketIOWithTimeout.java:407)
at
org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:322)
at
org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:203)
at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:407)
at
org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:304)
at
org.apache.hadoop.ipc.Client$Connection.access$1700(Client.java:176)
at org.apache.hadoop.ipc.Client.getConnection(Client.java:860)
at org.apache.hadoop.ipc.Client.call(Client.java:720)
... 13 more

Edson Ramiro


Re: java.io.IOException: Function not implemented

2010-03-29 Thread Todd Lipcon
Hi Edson,

What operating system are you on? What kernel version?

Thanks
-Todd

On Mon, Mar 29, 2010 at 12:01 PM, Edson Ramiro erlfi...@gmail.com wrote:

 Hi all,

 I'm trying to install Hadoop on a cluster, but I'm getting this error.

 I'm using java version 1.6.0_17 and hadoop-0.20.1+169.56.tar.gz from
 Cloudera.

 Its running in a NFS home shared between the nodes and masters.

 The NameNode works well, but all nodes try to connect and fail.

 Any Idea ?

 Thanks in Advance.

 == logs/hadoop-ramiro-datanode-a05.log ==
 2010-03-29 15:56:00,168 INFO org.apache.hadoop.ipc.Client: Retrying connect
 to server: lcpad/192.168.1.51:9000. Already tried 0 time(s).
 2010-03-29 15:56:01,172 INFO org.apache.hadoop.ipc.Client: Retrying connect
 to server: lcpad/192.168.1.51:9000. Already tried 1 time(s).
 2010-03-29 15:56:02,176 INFO org.apache.hadoop.ipc.Client: Retrying connect
 to server: lcpad/192.168.1.51:9000. Already tried 2 time(s).
 2010-03-29 15:56:03,180 INFO org.apache.hadoop.ipc.Client: Retrying connect
 to server: lcpad/192.168.1.51:9000. Already tried 3 time(s).
 2010-03-29 15:56:04,184 INFO org.apache.hadoop.ipc.Client: Retrying connect
 to server: lcpad/192.168.1.51:9000. Already tried 4 time(s).
 2010-03-29 15:56:05,188 INFO org.apache.hadoop.ipc.Client: Retrying connect
 to server: lcpad/192.168.1.51:9000. Already tried 5 time(s).
 2010-03-29 15:56:06,192 INFO org.apache.hadoop.ipc.Client: Retrying connect
 to server: lcpad/192.168.1.51:9000. Already tried 6 time(s).
 2010-03-29 15:56:07,196 INFO org.apache.hadoop.ipc.Client: Retrying connect
 to server: lcpad/192.168.1.51:9000. Already tried 7 time(s).
 2010-03-29 15:56:08,200 INFO org.apache.hadoop.ipc.Client: Retrying connect
 to server: lcpad/192.168.1.51:9000. Already tried 8 time(s).
 2010-03-29 15:56:09,204 INFO org.apache.hadoop.ipc.Client: Retrying connect
 to server: lcpad/192.168.1.51:9000. Already tried 9 time(s).
 2010-03-29 15:56:09,204 ERROR
 org.apache.hadoop.hdfs.server.datanode.DataNode: java.io.IOException: Call
 to lcpad/192.168.1.51:9000 failed on local exception: java.io.IOException:
 Function not implemented
at org.apache.hadoop.ipc.Client.wrapException(Client.java:775)
at org.apache.hadoop.ipc.Client.call(Client.java:743)
at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:220)
at $Proxy4.getProtocolVersion(Unknown Source)
at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:359)
at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:346)
at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:383)
at org.apache.hadoop.ipc.RPC.waitForProxy(RPC.java:314)
at org.apache.hadoop.ipc.RPC.waitForProxy(RPC.java:291)
at

 org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:278)
at
 org.apache.hadoop.hdfs.server.datanode.DataNode.init(DataNode.java:225)
at

 org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:1309)
at

 org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:1264)
at

 org.apache.hadoop.hdfs.server.datanode.DataNode.createDataNode(DataNode.java:1272)
at
 org.apache.hadoop.hdfs.server.datanode.DataNode.main(DataNode.java:1394)
 Caused by: java.io.IOException: Function not implemented
at sun.nio.ch.EPollArrayWrapper.epollCreate(Native Method)
at sun.nio.ch.EPollArrayWrapper.init(EPollArrayWrapper.java:68)
at sun.nio.ch.EPollSelectorImpl.init(EPollSelectorImpl.java:52)
at

 sun.nio.ch.EPollSelectorProvider.openSelector(EPollSelectorProvider.java:18)
at

 org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.get(SocketIOWithTimeout.java:407)
at

 org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:322)
at

 org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:203)
at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:407)
at
 org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:304)
at
 org.apache.hadoop.ipc.Client$Connection.access$1700(Client.java:176)
at org.apache.hadoop.ipc.Client.getConnection(Client.java:860)
at org.apache.hadoop.ipc.Client.call(Client.java:720)
... 13 more

 Edson Ramiro




-- 
Todd Lipcon
Software Engineer, Cloudera


Re: java.io.IOException: Function not implemented

2010-03-29 Thread Edson Ramiro
I'm using

Linux h02 2.6.32.9 #2 SMP Sat Mar 6 19:09:13 BRT 2010 x86_64 GNU/Linux

ram...@h02:~/hadoop$ cat /etc/debian_version
squeeze/sid

Thanks for reply

Edson Ramiro


On 29 March 2010 16:56, Todd Lipcon t...@cloudera.com wrote:

 Hi Edson,

 What operating system are you on? What kernel version?

 Thanks
 -Todd

 On Mon, Mar 29, 2010 at 12:01 PM, Edson Ramiro erlfi...@gmail.com wrote:

  Hi all,
 
  I'm trying to install Hadoop on a cluster, but I'm getting this error.
 
  I'm using java version 1.6.0_17 and hadoop-0.20.1+169.56.tar.gz from
  Cloudera.
 
  Its running in a NFS home shared between the nodes and masters.
 
  The NameNode works well, but all nodes try to connect and fail.
 
  Any Idea ?
 
  Thanks in Advance.
 
  == logs/hadoop-ramiro-datanode-a05.log ==
  2010-03-29 15:56:00,168 INFO org.apache.hadoop.ipc.Client: Retrying
 connect
  to server: lcpad/192.168.1.51:9000. Already tried 0 time(s).
  2010-03-29 15:56:01,172 INFO org.apache.hadoop.ipc.Client: Retrying
 connect
  to server: lcpad/192.168.1.51:9000. Already tried 1 time(s).
  2010-03-29 15:56:02,176 INFO org.apache.hadoop.ipc.Client: Retrying
 connect
  to server: lcpad/192.168.1.51:9000. Already tried 2 time(s).
  2010-03-29 15:56:03,180 INFO org.apache.hadoop.ipc.Client: Retrying
 connect
  to server: lcpad/192.168.1.51:9000. Already tried 3 time(s).
  2010-03-29 15:56:04,184 INFO org.apache.hadoop.ipc.Client: Retrying
 connect
  to server: lcpad/192.168.1.51:9000. Already tried 4 time(s).
  2010-03-29 15:56:05,188 INFO org.apache.hadoop.ipc.Client: Retrying
 connect
  to server: lcpad/192.168.1.51:9000. Already tried 5 time(s).
  2010-03-29 15:56:06,192 INFO org.apache.hadoop.ipc.Client: Retrying
 connect
  to server: lcpad/192.168.1.51:9000. Already tried 6 time(s).
  2010-03-29 15:56:07,196 INFO org.apache.hadoop.ipc.Client: Retrying
 connect
  to server: lcpad/192.168.1.51:9000. Already tried 7 time(s).
  2010-03-29 15:56:08,200 INFO org.apache.hadoop.ipc.Client: Retrying
 connect
  to server: lcpad/192.168.1.51:9000. Already tried 8 time(s).
  2010-03-29 15:56:09,204 INFO org.apache.hadoop.ipc.Client: Retrying
 connect
  to server: lcpad/192.168.1.51:9000. Already tried 9 time(s).
  2010-03-29 15:56:09,204 ERROR
  org.apache.hadoop.hdfs.server.datanode.DataNode: java.io.IOException:
 Call
  to lcpad/192.168.1.51:9000 failed on local exception:
 java.io.IOException:
  Function not implemented
 at org.apache.hadoop.ipc.Client.wrapException(Client.java:775)
 at org.apache.hadoop.ipc.Client.call(Client.java:743)
 at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:220)
 at $Proxy4.getProtocolVersion(Unknown Source)
 at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:359)
 at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:346)
 at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:383)
 at org.apache.hadoop.ipc.RPC.waitForProxy(RPC.java:314)
 at org.apache.hadoop.ipc.RPC.waitForProxy(RPC.java:291)
 at
 
 
 org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:278)
 at
  org.apache.hadoop.hdfs.server.datanode.DataNode.init(DataNode.java:225)
 at
 
 
 org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:1309)
 at
 
 
 org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:1264)
 at
 
 
 org.apache.hadoop.hdfs.server.datanode.DataNode.createDataNode(DataNode.java:1272)
 at
  org.apache.hadoop.hdfs.server.datanode.DataNode.main(DataNode.java:1394)
  Caused by: java.io.IOException: Function not implemented
 at sun.nio.ch.EPollArrayWrapper.epollCreate(Native Method)
 at sun.nio.ch.EPollArrayWrapper.init(EPollArrayWrapper.java:68)
 at sun.nio.ch.EPollSelectorImpl.init(EPollSelectorImpl.java:52)
 at
 
 
 sun.nio.ch.EPollSelectorProvider.openSelector(EPollSelectorProvider.java:18)
 at
 
 
 org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.get(SocketIOWithTimeout.java:407)
 at
 
 
 org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:322)
 at
 
 
 org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:203)
 at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:407)
 at
  org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:304)
 at
  org.apache.hadoop.ipc.Client$Connection.access$1700(Client.java:176)
 at org.apache.hadoop.ipc.Client.getConnection(Client.java:860)
 at org.apache.hadoop.ipc.Client.call(Client.java:720)
 ... 13 more
 
  Edson Ramiro
 



 --
 Todd Lipcon
 Software Engineer, Cloudera



Re: java.io.IOException: Function not implemented

2010-03-29 Thread Edson Ramiro
I'm not involved with Debian community :(

ram...@h02:~/hadoop$ cat /proc/sys/fs/epoll/max_user_watches
3373957

and the Java is not the OpenSDK.
The version is:

ram...@lcpad:/usr/lib/jvm/java-6-sun$ java -version
java version 1.6.0_17
Java(TM) SE Runtime Environment (build 1.6.0_17-b04)
Java HotSpot(TM) 64-Bit Server VM (build 14.3-b01, mixed mode)

Edson Ramiro


On 29 March 2010 17:14, Todd Lipcon t...@cloudera.com wrote:

 Hi Edson,

 It looks like for some reason your kernel does not have epoll enabled. It's
 very strange, since your kernel is very recent (in fact, bleeding edge!)

 Can you check the contents of /proc/sys/fs/epoll/max_user_watches

 Are you involved with the Debian community? This sounds like a general Java
 bug. Can you also please verify that you're using the Sun JVM and not
 OpenJDK (the debian folks like OpenJDK but it has subtle issues with
 Hadoop)
 You'll have to add a non-free repository and install sun-java6-jdk

 -Todd

 On Mon, Mar 29, 2010 at 1:05 PM, Edson Ramiro erlfi...@gmail.com wrote:

  I'm using
 
  Linux h02 2.6.32.9 #2 SMP Sat Mar 6 19:09:13 BRT 2010 x86_64 GNU/Linux
 
  ram...@h02:~/hadoop$ cat /etc/debian_version
  squeeze/sid
 
  Thanks for reply
 
  Edson Ramiro
 
 
  On 29 March 2010 16:56, Todd Lipcon t...@cloudera.com wrote:
 
   Hi Edson,
  
   What operating system are you on? What kernel version?
  
   Thanks
   -Todd
  
   On Mon, Mar 29, 2010 at 12:01 PM, Edson Ramiro erlfi...@gmail.com
  wrote:
  
Hi all,
   
I'm trying to install Hadoop on a cluster, but I'm getting this
 error.
   
I'm using java version 1.6.0_17 and hadoop-0.20.1+169.56.tar.gz
 from
Cloudera.
   
Its running in a NFS home shared between the nodes and masters.
   
The NameNode works well, but all nodes try to connect and fail.
   
Any Idea ?
   
Thanks in Advance.
   
== logs/hadoop-ramiro-datanode-a05.log ==
2010-03-29 15:56:00,168 INFO org.apache.hadoop.ipc.Client: Retrying
   connect
to server: lcpad/192.168.1.51:9000. Already tried 0 time(s).
2010-03-29 15:56:01,172 INFO org.apache.hadoop.ipc.Client: Retrying
   connect
to server: lcpad/192.168.1.51:9000. Already tried 1 time(s).
2010-03-29 15:56:02,176 INFO org.apache.hadoop.ipc.Client: Retrying
   connect
to server: lcpad/192.168.1.51:9000. Already tried 2 time(s).
2010-03-29 15:56:03,180 INFO org.apache.hadoop.ipc.Client: Retrying
   connect
to server: lcpad/192.168.1.51:9000. Already tried 3 time(s).
2010-03-29 15:56:04,184 INFO org.apache.hadoop.ipc.Client: Retrying
   connect
to server: lcpad/192.168.1.51:9000. Already tried 4 time(s).
2010-03-29 15:56:05,188 INFO org.apache.hadoop.ipc.Client: Retrying
   connect
to server: lcpad/192.168.1.51:9000. Already tried 5 time(s).
2010-03-29 15:56:06,192 INFO org.apache.hadoop.ipc.Client: Retrying
   connect
to server: lcpad/192.168.1.51:9000. Already tried 6 time(s).
2010-03-29 15:56:07,196 INFO org.apache.hadoop.ipc.Client: Retrying
   connect
to server: lcpad/192.168.1.51:9000. Already tried 7 time(s).
2010-03-29 15:56:08,200 INFO org.apache.hadoop.ipc.Client: Retrying
   connect
to server: lcpad/192.168.1.51:9000. Already tried 8 time(s).
2010-03-29 15:56:09,204 INFO org.apache.hadoop.ipc.Client: Retrying
   connect
to server: lcpad/192.168.1.51:9000. Already tried 9 time(s).
2010-03-29 15:56:09,204 ERROR
org.apache.hadoop.hdfs.server.datanode.DataNode: java.io.IOException:
   Call
to lcpad/192.168.1.51:9000 failed on local exception:
   java.io.IOException:
Function not implemented
   at org.apache.hadoop.ipc.Client.wrapException(Client.java:775)
   at org.apache.hadoop.ipc.Client.call(Client.java:743)
   at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:220)
   at $Proxy4.getProtocolVersion(Unknown Source)
   at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:359)
   at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:346)
   at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:383)
   at org.apache.hadoop.ipc.RPC.waitForProxy(RPC.java:314)
   at org.apache.hadoop.ipc.RPC.waitForProxy(RPC.java:291)
   at
   
   
  
 
 org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:278)
   at
   
  org.apache.hadoop.hdfs.server.datanode.DataNode.init(DataNode.java:225)
   at
   
   
  
 
 org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:1309)
   at
   
   
  
 
 org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:1264)
   at
   
   
  
 
 org.apache.hadoop.hdfs.server.datanode.DataNode.createDataNode(DataNode.java:1272)
   at
   
  org.apache.hadoop.hdfs.server.datanode.DataNode.main(DataNode.java:1394)
Caused by: java.io.IOException: Function not implemented
   at sun.nio.ch.EPollArrayWrapper.epollCreate(Native Method

Re: java.io.IOException: Function not implemented

2010-03-29 Thread Todd Lipcon
Hey Edson,

Unfortunately I'm not sure what's going on here - for whatever reason, the
kernel isn't allowing Java NIO to use epoll, and thus the IPC framework from
Hadoop isn't working correctly. I don't think this is a hadoop specific bug.

Does this issue occur on all of the nodes?

-Todd

On Mon, Mar 29, 2010 at 2:26 PM, Edson Ramiro erlfi...@gmail.com wrote:

 I'm not involved with Debian community :(

 ram...@h02:~/hadoop$ cat /proc/sys/fs/epoll/max_user_watches
 3373957

 and the Java is not the OpenSDK.
 The version is:

 ram...@lcpad:/usr/lib/jvm/java-6-sun$ java -version
 java version 1.6.0_17
 Java(TM) SE Runtime Environment (build 1.6.0_17-b04)
 Java HotSpot(TM) 64-Bit Server VM (build 14.3-b01, mixed mode)

 Edson Ramiro


 On 29 March 2010 17:14, Todd Lipcon t...@cloudera.com wrote:

  Hi Edson,
 
  It looks like for some reason your kernel does not have epoll enabled.
 It's
  very strange, since your kernel is very recent (in fact, bleeding edge!)
 
  Can you check the contents of /proc/sys/fs/epoll/max_user_watches
 
  Are you involved with the Debian community? This sounds like a general
 Java
  bug. Can you also please verify that you're using the Sun JVM and not
  OpenJDK (the debian folks like OpenJDK but it has subtle issues with
  Hadoop)
  You'll have to add a non-free repository and install sun-java6-jdk
 
  -Todd
 
  On Mon, Mar 29, 2010 at 1:05 PM, Edson Ramiro erlfi...@gmail.com
 wrote:
 
   I'm using
  
   Linux h02 2.6.32.9 #2 SMP Sat Mar 6 19:09:13 BRT 2010 x86_64 GNU/Linux
  
   ram...@h02:~/hadoop$ cat /etc/debian_version
   squeeze/sid
  
   Thanks for reply
  
   Edson Ramiro
  
  
   On 29 March 2010 16:56, Todd Lipcon t...@cloudera.com wrote:
  
Hi Edson,
   
What operating system are you on? What kernel version?
   
Thanks
-Todd
   
On Mon, Mar 29, 2010 at 12:01 PM, Edson Ramiro erlfi...@gmail.com
   wrote:
   
 Hi all,

 I'm trying to install Hadoop on a cluster, but I'm getting this
  error.

 I'm using java version 1.6.0_17 and hadoop-0.20.1+169.56.tar.gz
  from
 Cloudera.

 Its running in a NFS home shared between the nodes and masters.

 The NameNode works well, but all nodes try to connect and fail.

 Any Idea ?

 Thanks in Advance.

 == logs/hadoop-ramiro-datanode-a05.log ==
 2010-03-29 15:56:00,168 INFO org.apache.hadoop.ipc.Client: Retrying
connect
 to server: lcpad/192.168.1.51:9000. Already tried 0 time(s).
 2010-03-29 15:56:01,172 INFO org.apache.hadoop.ipc.Client: Retrying
connect
 to server: lcpad/192.168.1.51:9000. Already tried 1 time(s).
 2010-03-29 15:56:02,176 INFO org.apache.hadoop.ipc.Client: Retrying
connect
 to server: lcpad/192.168.1.51:9000. Already tried 2 time(s).
 2010-03-29 15:56:03,180 INFO org.apache.hadoop.ipc.Client: Retrying
connect
 to server: lcpad/192.168.1.51:9000. Already tried 3 time(s).
 2010-03-29 15:56:04,184 INFO org.apache.hadoop.ipc.Client: Retrying
connect
 to server: lcpad/192.168.1.51:9000. Already tried 4 time(s).
 2010-03-29 15:56:05,188 INFO org.apache.hadoop.ipc.Client: Retrying
connect
 to server: lcpad/192.168.1.51:9000. Already tried 5 time(s).
 2010-03-29 15:56:06,192 INFO org.apache.hadoop.ipc.Client: Retrying
connect
 to server: lcpad/192.168.1.51:9000. Already tried 6 time(s).
 2010-03-29 15:56:07,196 INFO org.apache.hadoop.ipc.Client: Retrying
connect
 to server: lcpad/192.168.1.51:9000. Already tried 7 time(s).
 2010-03-29 15:56:08,200 INFO org.apache.hadoop.ipc.Client: Retrying
connect
 to server: lcpad/192.168.1.51:9000. Already tried 8 time(s).
 2010-03-29 15:56:09,204 INFO org.apache.hadoop.ipc.Client: Retrying
connect
 to server: lcpad/192.168.1.51:9000. Already tried 9 time(s).
 2010-03-29 15:56:09,204 ERROR
 org.apache.hadoop.hdfs.server.datanode.DataNode:
 java.io.IOException:
Call
 to lcpad/192.168.1.51:9000 failed on local exception:
java.io.IOException:
 Function not implemented
at
 org.apache.hadoop.ipc.Client.wrapException(Client.java:775)
at org.apache.hadoop.ipc.Client.call(Client.java:743)
at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:220)
at $Proxy4.getProtocolVersion(Unknown Source)
at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:359)
at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:346)
at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:383)
at org.apache.hadoop.ipc.RPC.waitForProxy(RPC.java:314)
at org.apache.hadoop.ipc.RPC.waitForProxy(RPC.java:291)
at


   
  
 
 org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:278)
at

  
 org.apache.hadoop.hdfs.server.datanode.DataNode.init(DataNode.java:225)
at


   
  
 
 org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance