[ https://issues.apache.org/jira/browse/HDFS-8469?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14556632#comment-14556632 ]
Colin Patrick McCabe commented on HDFS-8469: -------------------------------------------- It looks like this behavior was introduced by HDFS-5138. I skimmed the comments, but I didn't see any discussion of datanode lock file changes. As far as I can tell, the fact that locking is now disabled on the datanode was unintentional. [~atm], [~tlipcon], [~sureshms], any perspective on this? > Lockfiles are not being created for datanode storage directories > ---------------------------------------------------------------- > > Key: HDFS-8469 > URL: https://issues.apache.org/jira/browse/HDFS-8469 > Project: Hadoop HDFS > Issue Type: Bug > Components: datanode > Reporter: Colin Patrick McCabe > Assignee: Colin Patrick McCabe > > Lockfiles are not being created for datanode storage directories. Due to a > mixup, we are initializing the StorageDirectory class with shared=true (an > option which was only intended for NFS directories used to implement NameNode > HA). Setting shared=true disables lockfile generation and prints a log > message like this: > {code} > 2015-05-22 11:45:16,367 INFO common.Storage (Storage.java:lock(675)) - > Locking is disabled for > /home/cmccabe/hadoop2/hadoop-hdfs-project/hadoop-hdfs/target/ > test/data/dfs/data/data5/current/BP-122766180-127.0.0.1-1432320314834 > {code} > Without lock files, we could accidentally spawn two datanode processes using > the same directories without realizing it. -- This message was sent by Atlassian JIRA (v6.3.4#6332)