The data node logs are on the datanode machines in the log directory. You may wish to buy my book and read chapter 4 on hdfs management.
On Thu, May 14, 2009 at 9:39 PM, Pankil Doshi <forpan...@gmail.com> wrote: > Can u guide me where can I find datanode log files? As I cannot find it in > $hadoop/logs and so. > > I can only find following files in logs folder :- > > hadoop-hadoop-namenode-hadoopmaster.log > hadoop-hadoop-namenode-hadoopmaster.out > hadoop-hadoop-namenode-hadoopmaster.out.1 > hadoop-hadoop-secondarynamenode-hadoopmaster.log > hadoop-hadoop-secondarynamenode-hadoopmaster.out > hadoop-hadoop-secondarynamenode-hadoopmaster.out.1 > history > > > Thanks > Pankil > > On Thu, May 14, 2009 at 11:27 PM, jason hadoop <jason.had...@gmail.com > >wrote: > > > You have to examine the datanode log files > > the namenode does not start the datanodes, the start script does. > > The name node passively waits for the datanodes to connect to it. > > > > On Thu, May 14, 2009 at 6:43 PM, Pankil Doshi <forpan...@gmail.com> > wrote: > > > > > Hello Everyone, > > > > > > Actually I had a cluster which was up. > > > > > > But i stopped the cluster as i wanted to format it.But cant start it > > back. > > > > > > 1)when i give "start-dfs.sh" I get following on screen > > > > > > starting namenode, logging to > > > > /Hadoop/hadoop-0.18.3/bin/../logs/hadoop-hadoop-namenode-hadoopmaster.out > > > slave1.local: starting datanode, logging to > > > /Hadoop/hadoop-0.18.3/bin/../logs/hadoop-hadoop-datanode-Slave1.out > > > slave3.local: starting datanode, logging to > > > /Hadoop/hadoop-0.18.3/bin/../logs/hadoop-hadoop-datanode-Slave3.out > > > slave4.local: starting datanode, logging to > > > /Hadoop/hadoop-0.18.3/bin/../logs/hadoop-hadoop-datanode-Slave4.out > > > slave2.local: starting datanode, logging to > > > /Hadoop/hadoop-0.18.3/bin/../logs/hadoop-hadoop-datanode-Slave2.out > > > slave5.local: starting datanode, logging to > > > /Hadoop/hadoop-0.18.3/bin/../logs/hadoop-hadoop-datanode-Slave5.out > > > slave6.local: starting datanode, logging to > > > /Hadoop/hadoop-0.18.3/bin/../logs/hadoop-hadoop-datanode-Slave6.out > > > slave9.local: starting datanode, logging to > > > /Hadoop/hadoop-0.18.3/bin/../logs/hadoop-hadoop-datanode-Slave9.out > > > slave8.local: starting datanode, logging to > > > /Hadoop/hadoop-0.18.3/bin/../logs/hadoop-hadoop-datanode-Slave8.out > > > slave7.local: starting datanode, logging to > > > /Hadoop/hadoop-0.18.3/bin/../logs/hadoop-hadoop-datanode-Slave7.out > > > slave10.local: starting datanode, logging to > > > /Hadoop/hadoop-0.18.3/bin/../logs/hadoop-hadoop-datanode-Slave10.out > > > hadoopmaster.local: starting secondarynamenode, logging to > > > > > > > > > /Hadoop/hadoop-0.18.3/bin/../logs/hadoop-hadoop-secondarynamenode-hadoopmaster.out > > > > > > > > > 2) from log file named "hadoop-hadoop-namenode-hadoopmaster.log" I get > > > following > > > > > > > > > > > > 2009-05-14 20:28:23,515 INFO org.apache.hadoop.dfs.NameNode: > STARTUP_MSG: > > > /************************************************************ > > > STARTUP_MSG: Starting NameNode > > > STARTUP_MSG: host = hadoopmaster/127.0.0.1 > > > STARTUP_MSG: args = [] > > > STARTUP_MSG: version = 0.18.3 > > > STARTUP_MSG: build = > > > https://svn.apache.org/repos/asf/hadoop/core/branches/branch-0.18 -r > > > 736250; > > > compiled by 'ndaley' on Thu Jan 22 23:12:08 UTC 2009 > > > ************************************************************/ > > > 2009-05-14 20:28:23,717 INFO org.apache.hadoop.ipc.metrics.RpcMetrics: > > > Initializing RPC Metrics with hostName=NameNode, port=9000 > > > 2009-05-14 20:28:23,728 INFO org.apache.hadoop.dfs.NameNode: Namenode > up > > > at: > > > hadoopmaster.local/192.168.0.1:9000 > > > 2009-05-14 20:28:23,733 INFO org.apache.hadoop.metrics.jvm.JvmMetrics: > > > Initializing JVM Metrics with processName=NameNode, sessionId=null > > > 2009-05-14 20:28:23,743 INFO org.apache.hadoop.dfs.NameNodeMetrics: > > > Initializing NameNodeMeterics using context > > > object:org.apache.hadoop.metrics.spi.NullContext > > > 2009-05-14 20:28:23,856 INFO org.apache.hadoop.fs.FSNamesystem: > > > > > > > > > fsOwner=hadoop,hadoop,adm,dialout,fax,cdrom,floppy,tape,audio,dip,video,plugdev,fuse,lpadmin,admin,sambashare > > > 2009-05-14 20:28:23,856 INFO org.apache.hadoop.fs.FSNamesystem: > > > supergroup=supergroup > > > 2009-05-14 20:28:23,856 INFO org.apache.hadoop.fs.FSNamesystem: > > > isPermissionEnabled=true > > > 2009-05-14 20:28:23,883 INFO org.apache.hadoop.dfs.FSNamesystemMetrics: > > > Initializing FSNamesystemMeterics using context > > > object:org.apache.hadoop.metrics.spi.NullContext > > > 2009-05-14 20:28:23,885 INFO org.apache.hadoop.fs.FSNamesystem: > > Registered > > > FSNamesystemStatusMBean > > > 2009-05-14 20:28:23,964 INFO org.apache.hadoop.dfs.Storage: Number of > > files > > > = 1 > > > 2009-05-14 20:28:23,971 INFO org.apache.hadoop.dfs.Storage: Number of > > files > > > under construction = 0 > > > 2009-05-14 20:28:23,971 INFO org.apache.hadoop.dfs.Storage: Image file > of > > > size 80 loaded in 0 seconds. > > > 2009-05-14 20:28:23,972 INFO org.apache.hadoop.dfs.Storage: Edits file > > > edits > > > of size 4 edits # 0 loaded in 0 seconds. > > > 2009-05-14 20:28:23,974 INFO org.apache.hadoop.fs.FSNamesystem: > Finished > > > loading FSImage in 155 msecs > > > 2009-05-14 20:28:23,976 INFO org.apache.hadoop.fs.FSNamesystem: Total > > > number > > > of blocks = 0 > > > 2009-05-14 20:28:23,988 INFO org.apache.hadoop.fs.FSNamesystem: Number > of > > > invalid blocks = 0 > > > 2009-05-14 20:28:23,988 INFO org.apache.hadoop.fs.FSNamesystem: Number > of > > > under-replicated blocks = 0 > > > 2009-05-14 20:28:23,988 INFO org.apache.hadoop.fs.FSNamesystem: Number > of > > > over-replicated blocks = 0 > > > 2009-05-14 20:28:23,988 INFO org.apache.hadoop.dfs.StateChange: STATE* > > > Leaving safe mode after 0 secs. > > > *2009-05-14 20:28:23,989 INFO org.apache.hadoop.dfs.StateChange: STATE* > > > Network topology has 0 racks and 0 datanodes* > > > 2009-05-14 20:28:23,989 INFO org.apache.hadoop.dfs.StateChange: STATE* > > > UnderReplicatedBlocks has 0 blocks > > > 2009-05-14 20:28:29,128 INFO org.mortbay.util.Credential: Checking > > Resource > > > aliases > > > 2009-05-14 20:28:29,243 INFO org.mortbay.http.HttpServer: Version > > > Jetty/5.1.4 > > > 2009-05-14 20:28:29,244 INFO org.mortbay.util.Container: Started > > > HttpContext[/static,/static] > > > 2009-05-14 20:28:29,245 INFO org.mortbay.util.Container: Started > > > HttpContext[/logs,/logs] > > > 2009-05-14 20:28:29,750 INFO org.mortbay.util.Container: Started > > > org.mortbay.jetty.servlet.webapplicationhand...@7fcebc9f > > > 2009-05-14 20:28:29,838 INFO org.mortbay.util.Container: Started > > > WebApplicationContext[/,/] > > > 2009-05-14 20:28:29,843 INFO org.mortbay.http.SocketListener: Started > > > SocketListener on 0.0.0.0:50070 > > > 2009-05-14 20:28:29,843 INFO org.mortbay.util.Container: Started > > > org.mortbay.jetty.ser...@61acfa31 > > > 2009-05-14 20:28:29,843 INFO org.apache.hadoop.fs.FSNamesystem: > > Web-server > > > up at: 0.0.0.0:50070 > > > 2009-05-14 20:28:29,843 INFO org.apache.hadoop.ipc.Server: IPC Server > > > Responder: starting > > > 2009-05-14 20:28:29,844 INFO org.apache.hadoop.ipc.Server: IPC Server > > > listener on 9000: starting > > > 2009-05-14 20:28:29,865 INFO org.apache.hadoop.ipc.Server: IPC Server > > > handler 0 on 9000: starting > > > 2009-05-14 20:28:29,876 INFO org.apache.hadoop.ipc.Server: IPC Server > > > handler 1 on 9000: starting > > > 2009-05-14 20:28:29,877 INFO org.apache.hadoop.ipc.Server: IPC Server > > > handler 2 on 9000: starting > > > 2009-05-14 20:28:29,877 INFO org.apache.hadoop.ipc.Server: IPC Server > > > handler 3 on 9000: starting > > > 2009-05-14 20:28:29,878 INFO org.apache.hadoop.ipc.Server: IPC Server > > > handler 4 on 9000: starting > > > 2009-05-14 20:28:29,879 INFO org.apache.hadoop.ipc.Server: IPC Server > > > handler 5 on 9000: starting > > > 2009-05-14 20:28:29,879 INFO org.apache.hadoop.ipc.Server: IPC Server > > > handler 6 on 9000: starting > > > 2009-05-14 20:28:29,881 INFO org.apache.hadoop.ipc.Server: IPC Server > > > handler 7 on 9000: starting > > > 2009-05-14 20:28:29,881 INFO org.apache.hadoop.ipc.Server: IPC Server > > > handler 8 on 9000: starting > > > 2009-05-14 20:28:29,882 INFO org.apache.hadoop.ipc.Server: IPC Server > > > handler 9 on 9000: starting > > > 2009-05-14 20:33:35,774 INFO org.apache.hadoop.fs.FSNamesystem: Roll > Edit > > > Log from 192.168.0.1 > > > 2009-05-14 20:33:35,775 INFO org.apache.hadoop.fs.FSNamesystem: Number > of > > > transactions: 0 Total time for transactions(ms): 0 Number of syncs: 0 > > > SyncTimes(ms): 0 > > > 2009-05-14 20:33:36,310 INFO org.apache.hadoop.fs.FSNamesystem: Roll > > > FSImage > > > from 192.168.0.1 > > > 2009-05-14 20:33:36,311 INFO org.apache.hadoop.fs.FSNamesystem: Number > of > > > transactions: 0 Total time for transactions(ms): 0 Number of syncs: 0 > > > SyncTimes(ms): 0 > > > > > > > > > > > > 3) my hadoop-site.xml for refrence > > > > > > <?xml version="1.0"?> > > > <?xml-stylesheet type="text/xsl" href="configuration.xsl"?> > > > > > > <!-- Put site-specific property overrides in this file. --> > > > <configuration> > > > <property> > > > <name>fs.default.name</name> > > > <value>hdfs://hadoopmaster.local:9000</value> > > > </property> > > > <property> > > > <name>mapred.job.tracker</name> > > > <value>hadoopmaster.local:9001</value> > > > </property> > > > <property> > > > <name>dfs.replication</name> > > > <value>3</value> > > > </property> > > > <property> > > > <name>mapred.child.java.opts</name> > > > <value>-Xmx512m</value> > > > </property> > > > <property> > > > <name>hadoop.tmp.dir</name> > > > <value>/Hadoop/Temp</value> > > > <description>A base for other temporary directories.</description> > > > </property> > > > <property> > > > <name>dfs.data.dir</name> > > > <value>/Hadoop/Data,/data/Hadoop</value> > > > <description>Determines where on the local filesystem an DFS data node > > > should store its blocks. If this is a comma-delimited > > > list of directories, then data will be stored in all named > > > directories, typically on different devices. > > > Directories that do not exist are ignored. > > > </description> > > > </property> > > > </configuration> > > > > > > > > > the main thing i find in log it says "*2009-05-14 20:28:23,989 INFO > > > org.apache.hadoop.dfs.StateChange: STATE* Network topology has 0 racks > > and > > > 0 > > > datanodes" *which means it cannot start datanodes*.*but why is it so? I > > > have > > > all my datanodes in my slave.xml and it detects that on start up screen > > > when > > > i give start-dfs.sh. > > > > > > > > > Can anyone throw some lights on my problem. > > > > > > Thanks > > > Pankil > > > > > > > > > > > -- > > Alpha Chapters of my book on Hadoop are available > > http://www.apress.com/book/view/9781430219422 > > www.prohadoopbook.com a community for Hadoop Professionals > > > -- Alpha Chapters of my book on Hadoop are available http://www.apress.com/book/view/9781430219422 www.prohadoopbook.com a community for Hadoop Professionals