Hi, I have set-up an 8-node private hadoop cluster having IP-addresses as follows:
192.168.1.10 (master) 192.168.1.11 192.168.1.12 192.168.1.13 192.168.1.14 192.168.1.15 192.168.1.16 192.168.1.17 Address 192.168.1.10, master, is running NN+JT and all other nodes are slaves i.e. running DN+TT. I am trying to put data on HDFS using command: hadoop dfs -put 8GB_input 8GB_input I have noticed that some blocks are not replicated/placed on nodes with IP addresses 192.168.1.11, 192.168.1.15, and 192.168.1.16 and i get the following error messages: ------------------------------------ $ hadoop dfs -put 8GB_input 8GB_input 09/08/31 18:25:45 INFO hdfs.DFSClient: Exception in createBlockOutputStream java.io.IOException: Bad connect ack with firstBadLink 192.168.1.11:50010 09/08/31 18:25:45 INFO hdfs.DFSClient: Abandoning block blk_-8575812198227241296_1001 09/08/31 18:25:51 INFO hdfs.DFSClient: Exception in createBlockOutputStream java.io.IOException: Bad connect ack with firstBadLink 192.168.1.16:50010 09/08/31 18:25:51 INFO hdfs.DFSClient: Abandoning block blk_-2932256218448902464_1001 09/08/31 18:25:57 INFO hdfs.DFSClient: Exception in createBlockOutputStream java.io.IOException: Bad connect ack with firstBadLink 192.168.1.11:50010 09/08/31 18:25:57 INFO hdfs.DFSClient: Abandoning block blk_-1014449966480421244_1001 09/08/31 18:26:03 INFO hdfs.DFSClient: Exception in createBlockOutputStream java.io.IOException: Bad connect ack with firstBadLink 192.168.1.16:50010 09/08/31 18:26:03 INFO hdfs.DFSClient: Abandoning block blk_7193173823538206978_1001 09/08/31 18:26:09 WARN hdfs.DFSClient: DataStreamer Exception: java.io.IOException: Unable to create new block. at org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.nextBlockOutputStream(DFSClient.java:2731) at org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.access$2000(DFSClient.java:1996) at org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$DataStreamer.run(DFSClient.java:2182) 09/08/31 18:26:09 WARN hdfs.DFSClient: Error Recovery for block blk_7193173823538206978_1001 bad datanode[2] nodes == null 09/08/31 18:26:09 WARN hdfs.DFSClient: Could not get block locations. Source file "/user/umer/8GB_input" - Aborting... put: Bad connect ack with firstBadLink 192.168.1.16:50010 ------------------------------------------------- Sometimes the input file is replicated successfully (excluding these three nodes) and sometimes the copy process i.e. 'hdfs -put input input' terminates. NOTE: My replication factor = 3. I am able to see that all machines are up-and-running using Web-UI http://192.168.1.10:50070. I will be greatful for any suggestion/comment in this regard. thanks, --umer _________________________________________________________________ Share your memories online with anyone you want. http://www.microsoft.com/middleeast/windows/windowslive/products/photos-share.aspx?tab=1