[ 
https://issues.apache.org/jira/browse/HDFS-12590?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dhaval Patel updated HDFS-12590:
--------------------------------
    Affects Version/s: 2.7.3
          Environment: RHEL7
          Description: 
{code:java}
2017-10-02 14:04:44,862 INFO  datanode.DataNode (BPServiceActor.java:run(733)) 
- Block pool <registering> (Datanode Uuid unassigned) service to 
master5.xxxxxx.local/10.10.10.10:8020 starting to offer serv
ice
2017-10-02 14:04:44,867 INFO  ipc.Server (Server.java:run(1045)) - IPC Server 
Responder: starting
2017-10-02 14:04:44,867 INFO  ipc.Server (Server.java:run(881)) - IPC Server 
listener on 8010: starting
2017-10-02 14:04:45,066 INFO  common.Storage 
(DataStorage.java:getParallelVolumeLoadThreadsNum(384)) - Using 2 threads to 
upgrade data directories (dfs.datanode.parallel.volumes.load.threads.num=2, 
dataDirs=
2)
2017-10-03 14:06:10,525 ERROR common.Storage (Storage.java:tryLock(783)) - 
Failed to acquire lock on /data1/hadoop/hdfs/data/in_use.lock. If this storage 
directory is mounted via NFS, ensure that the appropr
iate nfs lock services are running.
java.io.IOException: Resource temporarily unavailable
        at java.io.RandomAccessFile.writeBytes(Native Method)
        at java.io.RandomAccessFile.write(RandomAccessFile.java:512)
        at 
org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.tryLock(Storage.java:773)
        at 
org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.lock(Storage.java:736)
        at 
org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.analyzeStorage(Storage.java:549)
        at 
org.apache.hadoop.hdfs.server.datanode.DataStorage.loadStorageDirectory(DataStorage.java:299)
        at 
org.apache.hadoop.hdfs.server.datanode.DataStorage.loadDataStorage(DataStorage.java:438)
        at 
org.apache.hadoop.hdfs.server.datanode.DataStorage.addStorageLocations(DataStorage.java:417)
        at 
org.apache.hadoop.hdfs.server.datanode.DataStorage.recoverTransitionRead(DataStorage.java:595)
        at 
org.apache.hadoop.hdfs.server.datanode.DataNode.initStorage(DataNode.java:1483)
        at 
org.apache.hadoop.hdfs.server.datanode.DataNode.initBlockPool(DataNode.java:1448)
        at 
org.apache.hadoop.hdfs.server.datanode.BPOfferService.verifyAndSetNamespaceInfo(BPOfferService.java:319)
        at 
org.apache.hadoop.hdfs.server.datanode.BPServiceActor.connectToNNAndHandshake(BPServiceActor.java:267)
        at 
org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:740)
        at java.lang.Thread.run(Thread.java:745)
2017-10-03 14:06:10,542 WARN  common.Storage 
(DataStorage.java:loadDataStorage(449)) - Failed to add storage directory 
[DISK]file:/data1/hadoop/hdfs/data/
java.io.IOException: Resource temporarily unavailable
        at java.io.RandomAccessFile.writeBytes(Native Method)
        at java.io.RandomAccessFile.write(RandomAccessFile.java:512)
        at 
org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.tryLock(Storage.java:773)
        at 
org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.lock(Storage.java:736)
        at 
org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.analyzeStorage(Storage.java:549)
        at 
org.apache.hadoop.hdfs.server.datanode.DataStorage.loadStorageDirectory(DataStorage.java:299)
        at 
org.apache.hadoop.hdfs.server.datanode.DataStorage.loadDataStorage(DataStorage.java:438)
        at 
org.apache.hadoop.hdfs.server.datanode.DataStorage.addStorageLocations(DataStorage.java:417)
        at 
org.apache.hadoop.hdfs.server.datanode.DataStorage.recoverTransitionRead(DataStorage.java:595)
        at 
org.apache.hadoop.hdfs.server.datanode.DataNode.initStorage(DataNode.java:1483)
        at 
org.apache.hadoop.hdfs.server.datanode.DataNode.initBlockPool(DataNode.java:1448)
        at 
org.apache.hadoop.hdfs.server.datanode.BPOfferService.verifyAndSetNamespaceInfo(BPOfferService.java:319)
        at 
org.apache.hadoop.hdfs.server.datanode.BPServiceActor.connectToNNAndHandshake(BPServiceActor.java:267)
        at 
org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:740)
        at java.lang.Thread.run(Thread.java:745)
2017-10-03 18:03:16,928 ERROR datanode.DataNode (LogAdapter.java:error(71)) - 
RECEIVED SIGNAL 15: SIGTERM
2017-10-03 18:03:16,934 INFO  datanode.DataNode (LogAdapter.java:info(47)) - 
SHUTDOWN_MSG:
/************************************************************
SHUTDOWN_MSG: Shutting down DataNode at dc-slave29.xxxxxx.local/10.10.10.10
************************************************************/
2017-10-03 18:03:23,093 INFO  datanode.DataNode (LogAdapter.java:info(47)) - 
STARTUP_MSG:
/************************************************************
STARTUP_MSG: Starting DataNode
STARTUP_MSG:   user = hdfs
STARTUP_MSG:   host = xx-slave29.xxxxxx.local/10.10.10.10
STARTUP_MSG:   args = []
STARTUP_MSG:   version = 2.7.3.2.5.3.0-37
 
{code}


  was:

{code:java}
2017-10-02 14:04:44,862 INFO  datanode.DataNode (BPServiceActor.java:run(733)) 
- Block pool <registering> (Datanode Uuid unassigned) service to 
master5.xxxxxx.local/10.10.10.10:8020 starting to offer serv
ice
2017-10-02 14:04:44,867 INFO  ipc.Server (Server.java:run(1045)) - IPC Server 
Responder: starting
2017-10-02 14:04:44,867 INFO  ipc.Server (Server.java:run(881)) - IPC Server 
listener on 8010: starting
2017-10-02 14:04:45,066 INFO  common.Storage 
(DataStorage.java:getParallelVolumeLoadThreadsNum(384)) - Using 2 threads to 
upgrade data directories (dfs.datanode.parallel.volumes.load.threads.num=2, 
dataDirs=
2)
2017-10-03 14:06:10,525 ERROR common.Storage (Storage.java:tryLock(783)) - 
Failed to acquire lock on /data1/hadoop/hdfs/data/in_use.lock. If this storage 
directory is mounted via NFS, ensure that the appropr
iate nfs lock services are running.
java.io.IOException: Resource temporarily unavailable
        at java.io.RandomAccessFile.writeBytes(Native Method)
        at java.io.RandomAccessFile.write(RandomAccessFile.java:512)
        at 
org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.tryLock(Storage.java:773)
        at 
org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.lock(Storage.java:736)
        at 
org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.analyzeStorage(Storage.java:549)
        at 
org.apache.hadoop.hdfs.server.datanode.DataStorage.loadStorageDirectory(DataStorage.java:299)
        at 
org.apache.hadoop.hdfs.server.datanode.DataStorage.loadDataStorage(DataStorage.java:438)
        at 
org.apache.hadoop.hdfs.server.datanode.DataStorage.addStorageLocations(DataStorage.java:417)
        at 
org.apache.hadoop.hdfs.server.datanode.DataStorage.recoverTransitionRead(DataStorage.java:595)
        at 
org.apache.hadoop.hdfs.server.datanode.DataNode.initStorage(DataNode.java:1483)
        at 
org.apache.hadoop.hdfs.server.datanode.DataNode.initBlockPool(DataNode.java:1448)
        at 
org.apache.hadoop.hdfs.server.datanode.BPOfferService.verifyAndSetNamespaceInfo(BPOfferService.java:319)
        at 
org.apache.hadoop.hdfs.server.datanode.BPServiceActor.connectToNNAndHandshake(BPServiceActor.java:267)
        at 
org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:740)
        at java.lang.Thread.run(Thread.java:745)
2017-10-03 14:06:10,542 WARN  common.Storage 
(DataStorage.java:loadDataStorage(449)) - Failed to add storage directory 
[DISK]file:/data1/hadoop/hdfs/data/
java.io.IOException: Resource temporarily unavailable
        at java.io.RandomAccessFile.writeBytes(Native Method)
        at java.io.RandomAccessFile.write(RandomAccessFile.java:512)
        at 
org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.tryLock(Storage.java:773)
        at 
org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.lock(Storage.java:736)
        at 
org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.analyzeStorage(Storage.java:549)
        at 
org.apache.hadoop.hdfs.server.datanode.DataStorage.loadStorageDirectory(DataStorage.java:299)
        at 
org.apache.hadoop.hdfs.server.datanode.DataStorage.loadDataStorage(DataStorage.java:438)
        at 
org.apache.hadoop.hdfs.server.datanode.DataStorage.addStorageLocations(DataStorage.java:417)
        at 
org.apache.hadoop.hdfs.server.datanode.DataStorage.recoverTransitionRead(DataStorage.java:595)
        at 
org.apache.hadoop.hdfs.server.datanode.DataNode.initStorage(DataNode.java:1483)
        at 
org.apache.hadoop.hdfs.server.datanode.DataNode.initBlockPool(DataNode.java:1448)
        at 
org.apache.hadoop.hdfs.server.datanode.BPOfferService.verifyAndSetNamespaceInfo(BPOfferService.java:319)
        at 
org.apache.hadoop.hdfs.server.datanode.BPServiceActor.connectToNNAndHandshake(BPServiceActor.java:267)
        at 
org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:740)
        at java.lang.Thread.run(Thread.java:745)
2017-10-03 18:03:16,928 ERROR datanode.DataNode (LogAdapter.java:error(71)) - 
RECEIVED SIGNAL 15: SIGTERM
2017-10-03 18:03:16,934 INFO  datanode.DataNode (LogAdapter.java:info(47)) - 
SHUTDOWN_MSG:
/************************************************************
SHUTDOWN_MSG: Shutting down DataNode at dc-slave29.xxxxxx.local/10.10.10.10
************************************************************/
2017-10-03 18:03:23,093 INFO  datanode.DataNode (LogAdapter.java:info(47)) - 
STARTUP_MSG:
/************************************************************
STARTUP_MSG: Starting DataNode
STARTUP_MSG:   user = hdfs
STARTUP_MSG:   host = xx-slave29.xxxxxx.local/10.10.10.10
STARTUP_MSG:   args = []
STARTUP_MSG:   version = 2.7.3.2.5.3.0-37
 
{code}


          Component/s: datanode

> datanode process running in dead state for over 24 hours
> --------------------------------------------------------
>
>                 Key: HDFS-12590
>                 URL: https://issues.apache.org/jira/browse/HDFS-12590
>             Project: Hadoop HDFS
>          Issue Type: Bug
>          Components: datanode
>    Affects Versions: 2.7.3
>         Environment: RHEL7
>            Reporter: Dhaval Patel
>
> {code:java}
> 2017-10-02 14:04:44,862 INFO  datanode.DataNode 
> (BPServiceActor.java:run(733)) - Block pool <registering> (Datanode Uuid 
> unassigned) service to master5.xxxxxx.local/10.10.10.10:8020 starting to 
> offer serv
> ice
> 2017-10-02 14:04:44,867 INFO  ipc.Server (Server.java:run(1045)) - IPC Server 
> Responder: starting
> 2017-10-02 14:04:44,867 INFO  ipc.Server (Server.java:run(881)) - IPC Server 
> listener on 8010: starting
> 2017-10-02 14:04:45,066 INFO  common.Storage 
> (DataStorage.java:getParallelVolumeLoadThreadsNum(384)) - Using 2 threads to 
> upgrade data directories (dfs.datanode.parallel.volumes.load.threads.num=2, 
> dataDirs=
> 2)
> 2017-10-03 14:06:10,525 ERROR common.Storage (Storage.java:tryLock(783)) - 
> Failed to acquire lock on /data1/hadoop/hdfs/data/in_use.lock. If this 
> storage directory is mounted via NFS, ensure that the appropr
> iate nfs lock services are running.
> java.io.IOException: Resource temporarily unavailable
>         at java.io.RandomAccessFile.writeBytes(Native Method)
>         at java.io.RandomAccessFile.write(RandomAccessFile.java:512)
>         at 
> org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.tryLock(Storage.java:773)
>         at 
> org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.lock(Storage.java:736)
>         at 
> org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.analyzeStorage(Storage.java:549)
>         at 
> org.apache.hadoop.hdfs.server.datanode.DataStorage.loadStorageDirectory(DataStorage.java:299)
>         at 
> org.apache.hadoop.hdfs.server.datanode.DataStorage.loadDataStorage(DataStorage.java:438)
>         at 
> org.apache.hadoop.hdfs.server.datanode.DataStorage.addStorageLocations(DataStorage.java:417)
>         at 
> org.apache.hadoop.hdfs.server.datanode.DataStorage.recoverTransitionRead(DataStorage.java:595)
>         at 
> org.apache.hadoop.hdfs.server.datanode.DataNode.initStorage(DataNode.java:1483)
>         at 
> org.apache.hadoop.hdfs.server.datanode.DataNode.initBlockPool(DataNode.java:1448)
>         at 
> org.apache.hadoop.hdfs.server.datanode.BPOfferService.verifyAndSetNamespaceInfo(BPOfferService.java:319)
>         at 
> org.apache.hadoop.hdfs.server.datanode.BPServiceActor.connectToNNAndHandshake(BPServiceActor.java:267)
>         at 
> org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:740)
>         at java.lang.Thread.run(Thread.java:745)
> 2017-10-03 14:06:10,542 WARN  common.Storage 
> (DataStorage.java:loadDataStorage(449)) - Failed to add storage directory 
> [DISK]file:/data1/hadoop/hdfs/data/
> java.io.IOException: Resource temporarily unavailable
>         at java.io.RandomAccessFile.writeBytes(Native Method)
>         at java.io.RandomAccessFile.write(RandomAccessFile.java:512)
>         at 
> org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.tryLock(Storage.java:773)
>         at 
> org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.lock(Storage.java:736)
>         at 
> org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.analyzeStorage(Storage.java:549)
>         at 
> org.apache.hadoop.hdfs.server.datanode.DataStorage.loadStorageDirectory(DataStorage.java:299)
>         at 
> org.apache.hadoop.hdfs.server.datanode.DataStorage.loadDataStorage(DataStorage.java:438)
>         at 
> org.apache.hadoop.hdfs.server.datanode.DataStorage.addStorageLocations(DataStorage.java:417)
>         at 
> org.apache.hadoop.hdfs.server.datanode.DataStorage.recoverTransitionRead(DataStorage.java:595)
>         at 
> org.apache.hadoop.hdfs.server.datanode.DataNode.initStorage(DataNode.java:1483)
>         at 
> org.apache.hadoop.hdfs.server.datanode.DataNode.initBlockPool(DataNode.java:1448)
>         at 
> org.apache.hadoop.hdfs.server.datanode.BPOfferService.verifyAndSetNamespaceInfo(BPOfferService.java:319)
>         at 
> org.apache.hadoop.hdfs.server.datanode.BPServiceActor.connectToNNAndHandshake(BPServiceActor.java:267)
>         at 
> org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:740)
>         at java.lang.Thread.run(Thread.java:745)
> 2017-10-03 18:03:16,928 ERROR datanode.DataNode (LogAdapter.java:error(71)) - 
> RECEIVED SIGNAL 15: SIGTERM
> 2017-10-03 18:03:16,934 INFO  datanode.DataNode (LogAdapter.java:info(47)) - 
> SHUTDOWN_MSG:
> /************************************************************
> SHUTDOWN_MSG: Shutting down DataNode at dc-slave29.xxxxxx.local/10.10.10.10
> ************************************************************/
> 2017-10-03 18:03:23,093 INFO  datanode.DataNode (LogAdapter.java:info(47)) - 
> STARTUP_MSG:
> /************************************************************
> STARTUP_MSG: Starting DataNode
> STARTUP_MSG:   user = hdfs
> STARTUP_MSG:   host = xx-slave29.xxxxxx.local/10.10.10.10
> STARTUP_MSG:   args = []
> STARTUP_MSG:   version = 2.7.3.2.5.3.0-37
>  
> {code}



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

---------------------------------------------------------------------
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org

Reply via email to