Hi Omprakash! What is your default replication set to? What kind of disks do your datanodes have? Were you able to start a cluster with a simple configuration before you started tuning it?
HDFS tries to create the default number of replicas for a block on different datanodes. The Namenode tries to give a list of datanodes that the client can write replicas of the block to. If the Namenode is not able to construct a list with adequate number of datanodes, you will see the message you are seeing. This may mean that datanodes are unhealthy (failed disks), or full (disks have no more space), being decomissioned ( HDFS will not write replicas on decomissioning datanodes) or misconfigured ( I'd suggest turning on storage classes only after a simple configuration works). When a client that was trying to write a file was killed (e.g. if you killed your MR job), after some time (hard limit expiring) the Namenode will try to recover the file. In your case the namenode is also not able to find enough datanodes for recovering the files. HTH Ravi On Tue, Jun 20, 2017 at 11:50 PM, omprakash <ompraka...@cdac.in> wrote: > Hi, > > > > I am receiving lots of *warning messages in namenodes* logs on ACTIVE NN > in my *HA Hadoop setup*. Below are the logs > > > > *“2017-06-21 12:11:26,523 WARN > org.apache.hadoop.hdfs.protocol.BlockStoragePolicy: Failed to place enough > replicas: expected size is 1 but only 0 storage types can be selected > (replication=2, selected=[], unavailable=[DISK], removed=[DISK], > policy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], > replicationFallbacks=[ARCHIVE]})* > > *2017-06-21 12:11:26,523 WARN > org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy: Failed > to place enough replicas, still in need of 1 to reach 2 > (unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, > storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, > newBlock=true) All required storage types are unavailable: > unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, > storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}* > > *2017-06-21 12:11:26,523 INFO org.apache.hadoop.hdfs.StateChange: BLOCK* > allocate blk_1073894332_153508, replicas=192.168.9.174:50010 > <http://192.168.9.174:50010> for /36962._COPYING_* > > *2017-06-21 12:11:26,810 INFO org.apache.hadoop.hdfs.StateChange: DIR* > completeFile: /36962._COPYING_ is closed by > DFSClient_NONMAPREDUCE_146762699_1* > > *2017-06-21 12:11:30,626 WARN > org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy: Failed > to place enough replicas, still in need of 1 to reach 2 > (unavailableStorages=[], storagePolicy=BlockStoragePolicy{HOT:7, > storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, > newBlock=true) For more information, please enable DEBUG log level on > org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy and > org.apache.hadoop.net <http://org.apache.hadoop.net>.NetworkTopology* > > *2017-06-21 12:11:30,626 WARN > org.apache.hadoop.hdfs.protocol.BlockStoragePolicy: Failed to place enough > replicas: expected size is 1 but only 0 storage types can be selected > (replication=2, selected=[], unavailable=[DISK], removed=[DISK], > policy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], > replicationFallbacks=[ARCHIVE]})”* > > > > I am also encountering exceptions in active namenode related to > LeaseManager > > > > *2017-06-21 12:13:16,706 INFO > org.apache.hadoop.hdfs.server.namenode.LeaseManager: [Lease. Holder: > DFSClient_NONMAPREDUCE_409197282_362092, pending creates: 1] has expired > hard limit* > > *2017-06-21 12:13:16,706 INFO > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Recovering [Lease. > Holder: DFSClient_NONMAPREDUCE_409197282_362092, pending creates: 1], > src=/user/hadoop/2106201707 > <(210)%20620-1707>/02d5adda-d90f-47cb-85d5-999a079f4d79* > > *2017-06-21 12:13:16,706 WARN org.apache.hadoop.hdfs.StateChange: DIR* > NameSystem.internalReleaseLease: Failed to release lease for file > /user/hadoop/2106201707 > <(210)%20620-1707>/02d5adda-d90f-47cb-85d5-999a079f4d79. Committed blocks > are waiting to be minimally replicated. Try again later.* > > *2017-06-21 12:13:16,706 ERROR > org.apache.hadoop.hdfs.server.namenode.LeaseManager: Cannot release the > path /user/hadoop/2106201707 > <(210)%20620-1707>/02d5adda-d90f-47cb-85d5-999a079f4d79 in the lease > [Lease. Holder: DFSClient_NONMAPREDUCE_409197282_362092, pending creates: > 1]* > > *org.apache.hadoop.hdfs.protocol.AlreadyBeingCreatedException: DIR* > NameSystem.internalReleaseLease: Failed to release lease for file > /user/hadoop/2106201707 > <(210)%20620-1707>/02d5adda-d90f-47cb-85d5-999a079f4d79. Committed blocks > are waiting to be minimally replicated. Try again later.* > > * at > org.apache.hadoop.hdfs.server.namenode.FSNamesystem.internalReleaseLease(FSNamesystem.java:3200)* > > * at > org.apache.hadoop.hdfs.server.namenode.LeaseManager.checkLeases(LeaseManager.java:383)* > > * at > org.apache.hadoop.hdfs.server.namenode.LeaseManager$Monitor.run(LeaseManager.java:329)* > > * at java.lang.Thread.run(Thread.java:745)* > > > > I have checked the two datanodes. Both are running and have enough space > for new data. > > > > *PS: I have 2 Namenode and 2 datanodes in Hadoop HA setup. The HA is > setuped using Qourom Journal Manager and Zookeeper server.* > > > > Any idea why these errors? > > > > *Regards* > > *Omprakash Paliwal* > > HPC-Medical and Bioinformatics Applications Group > > Centre for Development of Advanced Computing (C-DAC) > > Pune University campus, > > PUNE-411007 > > Maharashtra, India > > email:*ompraka...@cdac.in <ompraka...@cdac.in>* > > Contact : +91-20-25704231 <+91%2020%202570%204231> > > > > ------------------------------------------------------------ > ------------------------------------------------------------------- > [ C-DAC is on Social-Media too. Kindly follow us at: > Facebook: https://www.facebook.com/CDACINDIA & Twitter: @cdacindia ] > > This e-mail is for the sole use of the intended recipient(s) and may > contain confidential and privileged information. If you are not the > intended recipient, please contact the sender by reply e-mail and destroy > all copies and the original message. Any unauthorized review, use, > disclosure, dissemination, forwarding, printing or copying of this email > is strictly prohibited and appropriate legal action will be taken. > ------------------------------------------------------------ > ------------------------------------------------------------------- >