Thanks for your help. But unfortunatly this didn’t do the job. Here’s the 
Shellscript I’ve written to start my cluster (the scripts on the other node 
only contains the command to start the datanode respectively the command to 
start the Nodemanager on the other node (with the right user (hdfs / yarn)):

 

 

#!/bin/bash

# Start 
HDFS-------------------------------------------------------------------------------------------------------------------------

# Start Namenode

su - hdfs -c "$HADOOP_PREFIX/sbin/hadoop-daemon.sh --config $HADOOP_CONF_DIR 
--script hdfs start namenode"

wait

 

# Start all Datanodes

export HADOOP_SECURE_DN_USER=hdfs

su - hdfs -c "$HADOOP_PREFIX/sbin/hadoop-daemon.sh --config $HADOOP_CONF_DIR 
--script hdfs start datanode"

wait

ssh root@hadoop-data.klinger.local 'bash startDatanode.sh'

wait

 

# Start Resourcemanager

su - yarn -c "$HADOOP_YARN_HOME/sbin/yarn-daemon.sh --config $HADOOP_CONF_DIR 
start resourcemanager"

wait

 

# Start Nodemanager on all Nodes

su - yarn -c "$HADOOP_YARN_HOME/sbin/yarn-daemon.sh --config $HADOOP_CONF_DIR 
start nodemanager"

wait

ssh root@hadoop-data.klinger.local 'bash startNodemanager.sh'

wait

 

# Start Proxyserver

#su - yarn -c "$HADOOP_YARN_HOME/bin/yarn start proxyserver --config 
$HADOOP_CONF_DIR"

#wait

 

# Start Historyserver

su - mapred -c "$HADOOP_PREFIX/sbin/mr-jobhistory-daemon.sh start historyserver 
--config $HADOOP_CONF_DIR"

wait

 

This script generates the following output:

 

starting namenode, logging to 
/var/log/cluster/hadoop/hadoop-hdfs-namenode-hadoop.klinger.local.out

starting datanode, logging to 
/var/log/cluster/hadoop/hadoop-hdfs-datanode-hadoop.klinger.local.out

starting datanode, logging to 
/var/log/cluster/hadoop/hadoop-hdfs-datanode-hadoop-data.klinger.local.out

starting resourcemanager, logging to 
/var/log/cluster/yarn/yarn-yarn-resourcemanager-hadoop.klinger.local.out

starting nodemanager, logging to 
/var/log/cluster/yarn/yarn-yarn-nodemanager-hadoop.klinger.local.out

starting nodemanager, logging to 
/var/log/cluster/yarn/yarn-yarn-nodemanager-hadoop-data.klinger.local.out

starting historyserver, logging to 
/var/log/cluster/mapred/mapred-mapred-historyserver-hadoop.klinger.local.out

 

Following my stopscript and it’s output:

 

#!/bin/bash

# Stop 
HDFS------------------------------------------------------------------------------------------------

# Stop Namenode

su - hdfs -c "$HADOOP_PREFIX/sbin/hadoop-daemon.sh --config $HADOOP_CONF_DIR 
--script hdfs stop namenode"

 

# Stop all Datanodes

su - hdfs -c "$HADOOP_PREFIX/sbin/hadoop-daemon.sh --config $HADOOP_CONF_DIR 
--script hdfs stop datanode"

ssh root@hadoop-data.klinger.local 'bash stopDatanode.sh'

 

# Stop Resourcemanager

su - yarn -c "$HADOOP_YARN_HOME/sbin/yarn-daemon.sh --config $HADOOP_CONF_DIR 
stop resourcemanager"

 

#Stop Nodemanager on all Hosts

su - yarn -c "$HADOOP_YARN_HOME/sbin/yarn-daemon.sh --config $HADOOP_CONF_DIR 
stop nodemanager"

ssh root@hadoop-data.klinger.local 'bash stopNodemanager.sh'

 

#Stop Proxyserver

#su - yarn -c "$HADOOP_YARN_HOME/bin/yarn stop proxyserver --config 
$HADOOP_CONF_DIR"

 

#Stop Historyserver

su - mapred -c "$HADOOP_PREFIX/sbin/mr-jobhistory-daemon.sh stop historyserver 
--config $HADOOP_CONF_DIR"

 

stopping namenode

no datanode to stop

no datanode to stop

stopping resourcemanager

stopping nodemanager

stopping nodemanager

nodemanager did not stop gracefully after 5 seconds: killing with kill -9

stopping historyserver

 

Is there may be anything wrong with my commands?

 

Greets

DK

 

Von: Varun Kumar [mailto:varun....@gmail.com] 
Gesendet: Montag, 2. März 2015 05:28
An: user
Betreff: Re: Hadoop 2.6.0 - No DataNode to stop

 

1.Stop the service 

2.Change the permissions for log and pid directory once again to hdfs.

 

3.Start service with hdfs.

 

This will resolve the issue

 

On Sun, Mar 1, 2015 at 6:40 PM, Daniel Klinger <d...@web-computing.de 
<mailto:d...@web-computing.de> > wrote:

Thanks for your answer. 

 

I put the FQDN of the DataNodes in the slaves file on each node (one FQDN per 
line). Here’s the full DataNode log after the start (the log of the other 
DataNode is exactly the same):

 

2015-03-02 00:29:41,841 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: 
registered UNIX signal handlers for [TERM, HUP, INT]

2015-03-02 00:29:42,207 INFO org.apache.hadoop.metrics2.impl.MetricsConfig: 
loaded properties from hadoop-metrics2.properties

2015-03-02 00:29:42,312 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: 
Scheduled snapshot period at 10 second(s).

2015-03-02 00:29:42,313 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: 
DataNode metrics system started

2015-03-02 00:29:42,319 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: 
Configured hostname is hadoop.klinger.local

2015-03-02 00:29:42,327 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: 
Starting DataNode with maxLockedMemory = 0

2015-03-02 00:29:42,350 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: 
Opened streaming server at /0.0.0.0:50010 <http://0.0.0.0:50010> 

2015-03-02 00:29:42,357 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: 
Balancing bandwith is 1048576 bytes/s

2015-03-02 00:29:42,358 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: 
Number threads for balancing is 5

2015-03-02 00:29:42,458 INFO org.mortbay.log: Logging to 
org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via org.mortbay.log.Slf4jLog

2015-03-02 00:29:42,462 INFO org.apache.hadoop.http.HttpRequestLog: Http 
request log for http.requests.datanode is not defined

2015-03-02 00:29:42,474 INFO org.apache.hadoop.http.HttpServer2: Added global 
filter 'safety' (class=org.apache.hadoop.http.HttpServer2$QuotingInputFilter)

2015-03-02 00:29:42,476 INFO org.apache.hadoop.http.HttpServer2: Added filter 
static_user_filter 
(class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to 
context datanode

2015-03-02 00:29:42,476 INFO org.apache.hadoop.http.HttpServer2: Added filter 
static_user_filter 
(class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to 
context logs

2015-03-02 00:29:42,476 INFO org.apache.hadoop.http.HttpServer2: Added filter 
static_user_filter 
(class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to 
context static

2015-03-02 00:29:42,494 INFO org.apache.hadoop.http.HttpServer2: 
addJerseyResourcePackage: 
packageName=org.apache.hadoop.hdfs.server.datanode.web.resources;org.apache.hadoop.hdfs.web.resources,
 pathSpec=/webhdfs/v1/*

2015-03-02 00:29:42,499 INFO org.mortbay.log: jetty-6.1.26

2015-03-02 00:29:42,555 WARN org.mortbay.log: Can't reuse 
/tmp/Jetty_0_0_0_0_50075_datanode____hwtdwq, using 
/tmp/Jetty_0_0_0_0_50075_datanode____hwtdwq_3168831075162569402

2015-03-02 00:29:43,205 INFO org.mortbay.log: Started 
HttpServer2$SelectChannelConnectorWithSafeStartup@0.0.0.0:50075 
<http://SelectChannelConnectorWithSafeStartup@0.0.0.0:50075> 

2015-03-02 00:29:43,635 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: 
dnUserName = hdfs

2015-03-02 00:29:43,635 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: 
supergroup = supergroup

2015-03-02 00:29:43,802 INFO org.apache.hadoop.ipc.CallQueueManager: Using 
callQueue class java.util.concurrent.LinkedBlockingQueue

2015-03-02 00:29:43,823 INFO org.apache.hadoop.ipc.Server: Starting Socket 
Reader #1 for port 50020

2015-03-02 00:29:43,875 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: 
Opened IPC server at /0.0.0.0:50020 <http://0.0.0.0:50020> 

2015-03-02 00:29:43,913 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: 
Refresh request received for nameservices: null

2015-03-02 00:29:43,953 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: 
Starting BPOfferServices for nameservices: <default>

2015-03-02 00:29:43,973 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: 
Block pool <registering> (Datanode Uuid unassigned) service to 
hadoop.klinger.local/10.0.1.148:8020 <http://10.0.1.148:8020>  starting to 
offer service

2015-03-02 00:29:43,981 INFO org.apache.hadoop.ipc.Server: IPC Server 
Responder: starting

2015-03-02 00:29:43,982 INFO org.apache.hadoop.ipc.Server: IPC Server listener 
on 50020: starting

2015-03-02 00:29:44,620 INFO org.apache.hadoop.hdfs.server.common.Storage: 
DataNode version: -56 and NameNode layout version: -60

2015-03-02 00:29:44,641 INFO org.apache.hadoop.hdfs.server.common.Storage: Lock 
on /cluster/storage/datanode/in_use.lock acquired by nodename 
1660@hadoop.klinger.local <mailto:1660@hadoop.klinger.local> 

2015-03-02 00:29:44,822 INFO org.apache.hadoop.hdfs.server.common.Storage: 
Analyzing storage directories for bpid BP-158097147-10.0.1.148-1424966425688

2015-03-02 00:29:44,822 INFO org.apache.hadoop.hdfs.server.common.Storage: 
Locking is disabled

2015-03-02 00:29:44,825 INFO org.apache.hadoop.hdfs.server.common.Storage: 
Restored 0 block files from trash.

2015-03-02 00:29:44,829 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: 
Setting up storage: 
nsid=330980018;bpid=BP-158097147-10.0.1.148-1424966425688;lv=-56;nsInfo=lv=-60;cid=CID-a2c81934-b3ce-44aa-b920-436ee2f0d5a7;nsid=330980018;c=0;bpid=BP-158097147-10.0.1.148-1424966425688;dnuuid=a3b6c890-41ca-4bde-855c-015c67e6e0df

2015-03-02 00:29:44,996 INFO 
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl: Added new 
volume: /cluster/storage/datanode/current

2015-03-02 00:29:44,998 INFO 
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl: Added 
volume - /cluster/storage/datanode/current, StorageType: DISK

2015-03-02 00:29:45,035 INFO 
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl: Registered 
FSDatasetState MBean

2015-03-02 00:29:45,057 INFO 
org.apache.hadoop.hdfs.server.datanode.DirectoryScanner: Periodic Directory 
Tree Verification scan starting at 1425265856057 with interval 21600000

2015-03-02 00:29:45,064 INFO 
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl: Adding 
block pool BP-158097147-10.0.1.148-1424966425688

2015-03-02 00:29:45,071 INFO 
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl: Scanning 
block pool BP-158097147-10.0.1.148-1424966425688 on volume 
/cluster/storage/datanode/current...

2015-03-02 00:29:45,128 INFO 
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl: Time taken 
to scan block pool BP-158097147-10.0.1.148-1424966425688 on 
/cluster/storage/datanode/current: 56ms

2015-03-02 00:29:45,128 INFO 
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl: Total time 
to scan all replicas for block pool BP-158097147-10.0.1.148-1424966425688: 64ms

2015-03-02 00:29:45,128 INFO 
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl: Adding 
replicas to map for block pool BP-158097147-10.0.1.148-1424966425688 on volume 
/cluster/storage/datanode/current...

2015-03-02 00:29:45,129 INFO 
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl: Time to 
add replicas to map for block pool BP-158097147-10.0.1.148-1424966425688 on 
volume /cluster/storage/datanode/current: 0ms

2015-03-02 00:29:45,134 INFO 
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl: Total time 
to add all replicas to map: 5ms

2015-03-02 00:29:45,138 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: 
Block pool BP-158097147-10.0.1.148-1424966425688 (Datanode Uuid null) service 
to hadoop.klinger.local/10.0.1.148:8020 <http://10.0.1.148:8020>  beginning 
handshake with NN

2015-03-02 00:29:45,316 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: 
Block pool Block pool BP-158097147-10.0.1.148-1424966425688 (Datanode Uuid 
null) service to hadoop.klinger.local/10.0.1.148:8020 <http://10.0.1.148:8020>  
successfully registered with NN

2015-03-02 00:29:45,316 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: 
For namenode hadoop.klinger.local/10.0.1.148:8020 <http://10.0.1.148:8020>  
using DELETEREPORT_INTERVAL of 300000 msec  BLOCKREPORT_INTERVAL of 
21600000msec CACHEREPORT_INTERVAL of 10000msec Initial delay: 0msec; 
heartBeatInterval=3000

2015-03-02 00:29:45,751 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: 
Namenode Block pool BP-158097147-10.0.1.148-1424966425688 (Datanode Uuid 
a3b6c890-41ca-4bde-855c-015c67e6e0df) service to 
hadoop.klinger.local/10.0.1.148:8020 <http://10.0.1.148:8020>  trying to claim 
ACTIVE state with txid=24

2015-03-02 00:29:45,751 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: 
Acknowledging ACTIVE Namenode Block pool BP-158097147-10.0.1.148-1424966425688 
(Datanode Uuid a3b6c890-41ca-4bde-855c-015c67e6e0df) service to 
hadoop.klinger.local/10.0.1.148:8020 <http://10.0.1.148:8020> 

2015-03-02 00:29:45,883 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: 
Sent 1 blockreports 0 blocks total. Took 4 msec to generate and 126 msecs for 
RPC and NN processing.  Got back commands 
org.apache.hadoop.hdfs.server.protocol.FinalizeCommand@3d528774 
<mailto:org.apache.hadoop.hdfs.server.protocol.FinalizeCommand@3d528774> 

2015-03-02 00:29:45,883 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: 
Got finalize command for block pool BP-158097147-10.0.1.148-1424966425688

2015-03-02 00:29:45,891 INFO org.apache.hadoop.util.GSet: Computing capacity 
for map BlockMap

2015-03-02 00:29:45,891 INFO org.apache.hadoop.util.GSet: VM type       = 64-bit

2015-03-02 00:29:45,893 INFO org.apache.hadoop.util.GSet: 0.5% max memory 966.7 
MB = 4.8 MB

2015-03-02 00:29:45,893 INFO org.apache.hadoop.util.GSet: capacity      = 2^19 
= 524288 entries

2015-03-02 00:29:45,894 INFO 
org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Periodic Block 
Verification Scanner initialized with interval 504 hours for block pool 
BP-158097147-10.0.1.148-1424966425688

2015-03-02 00:29:45,900 INFO 
org.apache.hadoop.hdfs.server.datanode.DataBlockScanner: Added 
bpid=BP-158097147-10.0.1.148-1424966425688 to blockPoolScannerMap, new size=1

 

 

Dfsadmin –report (called as user hdfs on NameNode) generated following output. 
It looks like both DataNodes are available:

 

Configured Capacity: 985465716736 (917.79 GB)

Present Capacity: 929892360192 (866.03 GB)

DFS Remaining: 929892302848 (866.03 GB)

DFS Used: 57344 (56 KB)

DFS Used%: 0.00%

Under replicated blocks: 0

Blocks with corrupt replicas: 0

Missing blocks: 0

 

-------------------------------------------------

Live datanodes (2):

 

Name: 10.0.1.148:50010 <http://10.0.1.148:50010>  (hadoop.klinger.local)

Hostname: hadoop.klinger.local

Decommission Status : Normal

Configured Capacity: 492732858368 (458.89 GB)

DFS Used: 28672 (28 KB)

Non DFS Used: 27942051840 (26.02 GB)

DFS Remaining: 464790777856 (432.87 GB)

DFS Used%: 0.00%

DFS Remaining%: 94.33%

Configured Cache Capacity: 0 (0 B)

Cache Used: 0 (0 B)

Cache Remaining: 0 (0 B)

Cache Used%: 100.00%

Cache Remaining%: 0.00%

Xceivers: 1

Last contact: Mon Mar 02 00:38:00 CET 2015

 

 

Name: 10.0.1.89:50010 <http://10.0.1.89:50010>  (hadoop-data.klinger.local)

Hostname: hadoop-data.klinger.local

Decommission Status : Normal

Configured Capacity: 492732858368 (458.89 GB)

DFS Used: 28672 (28 KB)

Non DFS Used: 27631304704 (25.73 GB)

DFS Remaining: 465101524992 (433.16 GB)

DFS Used%: 0.00%

DFS Remaining%: 94.39%

Configured Cache Capacity: 0 (0 B)

Cache Used: 0 (0 B)

Cache Remaining: 0 (0 B)

Cache Used%: 100.00%

Cache Remaining%: 0.00%

Xceivers: 1

Last contact: Mon Mar 02 00:37:59 CET 2015

 

Any further thoughts?

 

Greets

DK

Von: Ulul [mailto:had...@ulul.org <mailto:had...@ulul.org> ] 
Gesendet: Sonntag, 1. März 2015 13:12


An: user@hadoop.apache.org <mailto:user@hadoop.apache.org> 
Betreff: Re: Hadoop 2.6.0 - No DataNode to stop

 

Hi

Did you check your slaves file is correct ?
That the datanode process is actually running ?
Did you check its log file ?
That the datanode is available ? (dfsadmin -report, through the WUI)

We need more detail

Ulul

Le 28/02/2015 22:05, Daniel Klinger a écrit :

Thanks but i know how to kill a process in Linux. But this didn’t answer the 
question why the command say no Datanode to stop instead of stopping the 
Datanode:
 
$HADOOP_PREFIX/sbin/hadoop-daemon.sh --config $HADOOP_CONF_DIR --script hdfs 
stop datanode
 
 

 

 

Von: Surbhi Gupta [mailto:surbhi.gupt...@gmail.com] 
Gesendet: Samstag, 28. Februar 2015 20:16
An: user@hadoop.apache.org <mailto:user@hadoop.apache.org> 
Betreff: Re: Hadoop 2.6.0 - No DataNode to stop

 

Issue jps and get the process id or 
Try to get the process id of datanode.

Issue ps-fu userid of the user through which datanode is running.

Then kill the process using kill -9

On 28 Feb 2015 09:38, "Daniel Klinger" <d...@web-computing.de 
<mailto:d...@web-computing.de> > wrote:

Hello,

 

I used a lot of Hadoop-Distributions. Now I’m trying to install a pure Hadoop 
on a little „cluster“ for testing (2 CentOS-VMs: 1 Name+DataNode 1 DataNode). I 
followed the instructions on the Documentation site: 
http://hadoop.apache.org/docs/current/hadoop-project-dist/hadoop-common/ClusterSetup.html.

 

I’m starting the Cluster like it is described in the Chapter „Operating the 
Hadoop Cluster“(with different users). The starting process works great. The 
PID-Files are created in /var/run and u can see that Folders and Files are 
created in the Data- and NameNode folders. I’m getting no errors in the 
log-files.

 

When I try to stop the cluster all Services are stopped (NameNode, 
ResourceManager etc.). But when I stop the DataNodes I’m getting the message: 
„No DataNode to stop“. The PID-File and the in_use.lock-File are still there 
and if I try to start the DataNode again I’m getting the error that the Process 
is already running. When I stop the DataNode as hdfs instead of root the PID 
and in_use-File are removed but I’m still getting the message: „No DataNode to 
stop“

 

What I’m doing wrong?

 

Greets

dk

 





 

-- 

Regards,

Varun Kumar.P

Reply via email to