Sean Chow created HDFS-14536:
--------------------------------

             Summary: It's safe to remove synchronized from reconcile scan
                 Key: HDFS-14536
                 URL: https://issues.apache.org/jira/browse/HDFS-14536
             Project: Hadoop HDFS
          Issue Type: Improvement
          Components: datanode
    Affects Versions: 2.7.0, 2.6.0
            Reporter: Sean Chow


I found all my my request blocked in every 6hours and use jstack to dump 
threads:
{code:java}
"1490900528@qtp-341796579-35" #132086969 daemon prio=5 os_prio=0 
tid=0x00007f7ed01c3800 nid=0x1a1c waiting for monitor entry [0x00007f7e9e42b000]
  java.lang.Thread.State: BLOCKED (on object monitor)
    at 
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsVolumeImpl.getDfsUsed(FsVolumeImpl.java:295)
 - waiting to lock <0x00000002c00410f8> (a 
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl)
 at 
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.getVolumeInfo(FsDatasetImpl.java:2323)

/// lots of BLOCKED for about one minute

"java.util.concurrent.ThreadPoolExecutor$Worker@49d8fff9[State = -1, empty 
queue]" #77 daemon prio=5 os_prio=0 tid=0x00007f7ed8590800 nid=0x49cc runnable 
[0x00007f7ea0344000]
  java.lang.Thread.State: RUNNABLE
    at 
org.apache.hadoop.hdfs.server.datanode.DatanodeUtil.idToBlockDir(DatanodeUtil.java:120)
    at 
org.apache.hadoop.hdfs.server.datanode.ReplicaInfo.getDir(ReplicaInfo.java:143)
    at 
org.apache.hadoop.hdfs.server.datanode.ReplicaInfo.<init>(ReplicaInfo.java:95)
    at 
org.apache.hadoop.hdfs.server.datanode.FinalizedReplica.<init>(FinalizedReplica.java:60)
    at 
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.getFinalizedBlocks(FsDatasetImpl.java:1510)
    - locked <0x00000002c00410f8> (a 
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl)
    at 
org.apache.hadoop.hdfs.server.datanode.DirectoryScanner.scan(DirectoryScanner.java:437)
    - locked <0x00000002c00410f8> (a 
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl)
    at 
org.apache.hadoop.hdfs.server.datanode.DirectoryScanner.reconcile(DirectoryScanner.java:404)
    at 
org.apache.hadoop.hdfs.server.datanode.DirectoryScanner.run(DirectoryScanner.java:360)
    at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
    at 
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
    at 
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
    at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
{code}
After reviewing the DirectoryScanner behavior, I think the lock 
`synchronized(dataset) in ` `scan()` can be removed safely.

 

As dataset.getFinalizedBlocks() is a synchronized method, the memReport will be 
generated from volumeMap with lock. After that , the outside lock 
`synchronized(dataset)` is needless but I've encounter the later while loop 
running 20 seconds

Yes, other blocks could be added or deleted from memory or disk, but it has no 
affect to get the differences between them. Because the later fixing processing 
will handle both sides.

 

 
{code:java}
  void scan() {
    clear();
    Map<String, ScanInfo[]> diskReport = getDiskReport();

    // Hold FSDataset lock to prevent further changes to the block map
    synchronized(dataset) {  // remove synchronized?
      for (Entry<String, ScanInfo[]> entry : diskReport.entrySet()) {
        String bpid = entry.getKey();
        ScanInfo[] blockpoolReport = entry.getValue();

        Stats statsRecord = new Stats(bpid);
        stats.put(bpid, statsRecord);
        LinkedList<ScanInfo> diffRecord = new LinkedList<ScanInfo>();
        diffs.put(bpid, diffRecord);

        statsRecord.totalBlocks = blockpoolReport.length;
        // getFinalizedBlocks is synchronized 
        List<FinalizedReplica> bl = dataset.getFinalizedBlocks(bpid);
        FinalizedReplica[] memReport = bl.toArray(new 
FinalizedReplica[bl.size()]);
        Arrays.sort(memReport); // Sort based on blockId

        int d = 0; // index for blockpoolReport
        int m = 0; // index for memReprot
        while (m < memReport.length && d < blockpoolReport.length) {
          // may take 20s to finish this loop
        }
      } //end for
    } //end synchronized
  }{code}
 

Let say block A is added and block B is deleted after 
`dataset.getFinalizedBlocks(bpid)` called. That means the disk  file and memory 
is updated. Because we have a memory finalized blocks snapshot before, and just 
care about those blocks consistence. 

If the difference report occasionally has that deleted block B, the later 
`checkAndUpdate` will handle it in the right way.

 

I've compile this patch and try will test it in my enviroment.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

---------------------------------------------------------------------
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org

Reply via email to