Hi Brian, Thanks for replying. I don't even know what the metrics really are so have never knowingly touched them. I have only done the basic core and site configuration along with master / slaves list and then a namenode format and started the cluster to get this error. All based on the current 0.20.1 download on apache.
I think I am using the NullContext looking at the properties file... could you please verify this is correct? Many thanks for your help! Tim # Configuration of the "dfs" context for null dfs.class=org.apache.hadoop.metrics.spi.NullContext # Configuration of the "dfs" context for file #dfs.class=org.apache.hadoop.metrics.file.FileContext #dfs.period=10 #dfs.fileName=/tmp/dfsmetrics.log # Configuration of the "dfs" context for ganglia # dfs.class=org.apache.hadoop.metrics.ganglia.GangliaContext # dfs.period=10 # dfs.servers=localhost:8649 # Configuration of the "mapred" context for null mapred.class=org.apache.hadoop.metrics.spi.NullContext # Configuration of the "mapred" context for file #mapred.class=org.apache.hadoop.metrics.file.FileContext #mapred.period=10 #mapred.fileName=/tmp/mrmetrics.log # Configuration of the "mapred" context for ganglia # mapred.class=org.apache.hadoop.metrics.ganglia.GangliaContext # mapred.period=10 # mapred.servers=localhost:8649 # Configuration of the "jvm" context for null jvm.class=org.apache.hadoop.metrics.spi.NullContext # Configuration of the "jvm" context for file #jvm.class=org.apache.hadoop.metrics.file.FileContext #jvm.period=10 #jvm.fileName=/tmp/jvmmetrics.log # Configuration of the "jvm" context for ganglia # jvm.class=org.apache.hadoop.metrics.ganglia.GangliaContext # jvm.period=10 # jvm.servers=localhost:8649 On Wed, Oct 14, 2009 at 3:12 PM, Brian Bockelman <bbock...@cse.unl.edu> wrote: > Hey Tim, > > Can you see if that goes away if you use the null metrics context (the error > appears to be coming from metrics - have you touched them?)? > > Brian > > On Oct 14, 2009, at 4:33 AM, tim robertson wrote: > >> Hi all, >> >> Using hadoop 0.20.1 I am seeing the following on my namenode startup. >> >> 2009-10-14 11:09:54,232 INFO org.apache.hadoop.ipc.Server: Error >> register getProtocolVersion >> java.lang.IllegalArgumentException: Duplicate >> metricsName:getProtocolVersion >> >> Could someone please point me in the right direction for diagnosing >> where I have gone wrong? >> >> Thanks >> Tim >> >> >> >> /************************************************************ >> STARTUP_MSG: Starting NameNode >> STARTUP_MSG: host = cluma.gbif.org/192.38.28.77 >> STARTUP_MSG: args = [] >> STARTUP_MSG: version = 0.20.1 >> STARTUP_MSG: build = >> http://svn.apache.org/repos/asf/hadoop/common/tags/release-0.20.1-rc1 >> -r 810220; compiled by 'oom' on Tue Sep 1 20:55:56 UTC 2009 >> ************************************************************/ >> 2009-10-14 11:09:53,010 INFO org.apache.hadoop.ipc.metrics.RpcMetrics: >> Initializing RPC Metrics with hostName=NameNode, port=8020 >> 2009-10-14 11:09:53,013 INFO >> org.apache.hadoop.hdfs.server.namenode.NameNode: Namenode up at: >> cluma.gbif.clu/192.168.76.254:8020 >> 2009-10-14 11:09:53,015 INFO org.apache.hadoop.metrics.jvm.JvmMetrics: >> Initializing JVM Metrics with processName=NameNode, sessionId=null >> 2009-10-14 11:09:53,019 INFO >> org.apache.hadoop.hdfs.server.namenode.metrics.NameNodeMetrics: >> Initializing NameNodeMeterics using context >> object:org.apache.hadoop.metrics.spi.NullContext >> 2009-10-14 11:09:53,056 INFO >> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: >> fsOwner=root,root,bin,daemon,sys,adm,disk,wheel >> 2009-10-14 11:09:53,057 INFO >> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: >> supergroup=supergroup >> 2009-10-14 11:09:53,057 INFO >> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: >> isPermissionEnabled=true >> 2009-10-14 11:09:53,064 INFO >> org.apache.hadoop.hdfs.server.namenode.metrics.FSNamesystemMetrics: >> Initializing FSNamesystemMetrics using context >> object:org.apache.hadoop.metrics.spi.NullContext >> 2009-10-14 11:09:53,065 INFO >> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Registered >> FSNamesystemStatusMBean >> 2009-10-14 11:09:53,090 INFO >> org.apache.hadoop.hdfs.server.common.Storage: Number of files = 1 >> 2009-10-14 11:09:53,094 INFO >> org.apache.hadoop.hdfs.server.common.Storage: Number of files under >> construction = 0 >> 2009-10-14 11:09:53,094 INFO >> org.apache.hadoop.hdfs.server.common.Storage: Image file of size 94 >> loaded in 0 seconds. >> 2009-10-14 11:09:53,094 INFO >> org.apache.hadoop.hdfs.server.common.Storage: Edits file >> /hadoop/name/current/edits of size 4 edits # 0 loaded in 0 seconds. >> 2009-10-14 11:09:53,098 INFO >> org.apache.hadoop.hdfs.server.common.Storage: Image file of size 94 >> saved in 0 seconds. >> 2009-10-14 11:09:53,113 INFO >> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Finished loading >> FSImage in 72 msecs >> 2009-10-14 11:09:53,114 INFO >> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Total number of >> blocks = 0 >> 2009-10-14 11:09:53,114 INFO >> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of invalid >> blocks = 0 >> 2009-10-14 11:09:53,114 INFO >> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of >> under-replicated blocks = 0 >> 2009-10-14 11:09:53,114 INFO >> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of >> over-replicated blocks = 0 >> 2009-10-14 11:09:53,114 INFO org.apache.hadoop.hdfs.StateChange: >> STATE* Leaving safe mode after 0 secs. >> 2009-10-14 11:09:53,114 INFO org.apache.hadoop.hdfs.StateChange: >> STATE* Network topology has 0 racks and 0 datanodes >> 2009-10-14 11:09:53,114 INFO org.apache.hadoop.hdfs.StateChange: >> STATE* UnderReplicatedBlocks has 0 blocks >> 2009-10-14 11:09:53,198 INFO org.mortbay.log: Logging to >> org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via >> org.mortbay.log.Slf4jLog >> 2009-10-14 11:09:53,242 INFO org.apache.hadoop.http.HttpServer: Port >> returned by webServer.getConnectors()[0].getLocalPort() before open() >> is -1. Opening the listener on 50070 >> 2009-10-14 11:09:53,243 INFO org.apache.hadoop.http.HttpServer: >> listener.getLocalPort() returned 50070 >> webServer.getConnectors()[0].getLocalPort() returned 50070 >> 2009-10-14 11:09:53,243 INFO org.apache.hadoop.http.HttpServer: Jetty >> bound to port 50070 >> 2009-10-14 11:09:53,243 INFO org.mortbay.log: jetty-6.1.14 >> 2009-10-14 11:09:53,433 INFO org.mortbay.log: Started >> selectchannelconnec...@0.0.0.0:50070 >> 2009-10-14 11:09:53,434 INFO >> org.apache.hadoop.hdfs.server.namenode.NameNode: Web-server up at: >> 0.0.0.0:50070 >> 2009-10-14 11:09:53,434 INFO org.apache.hadoop.ipc.Server: IPC Server >> Responder: starting >> 2009-10-14 11:09:53,435 INFO org.apache.hadoop.ipc.Server: IPC Server >> listener on 8020: starting >> 2009-10-14 11:09:53,438 INFO org.apache.hadoop.ipc.Server: IPC Server >> handler 0 on 8020: starting >> 2009-10-14 11:09:53,438 INFO org.apache.hadoop.ipc.Server: IPC Server >> handler 1 on 8020: starting >> 2009-10-14 11:09:53,438 INFO org.apache.hadoop.ipc.Server: IPC Server >> handler 2 on 8020: starting >> 2009-10-14 11:09:53,438 INFO org.apache.hadoop.ipc.Server: IPC Server >> handler 3 on 8020: starting >> 2009-10-14 11:09:53,438 INFO org.apache.hadoop.ipc.Server: IPC Server >> handler 4 on 8020: starting >> 2009-10-14 11:09:53,439 INFO org.apache.hadoop.ipc.Server: IPC Server >> handler 6 on 8020: starting >> 2009-10-14 11:09:53,439 INFO org.apache.hadoop.ipc.Server: IPC Server >> handler 7 on 8020: starting >> 2009-10-14 11:09:53,439 INFO org.apache.hadoop.ipc.Server: IPC Server >> handler 8 on 8020: starting >> 2009-10-14 11:09:53,441 INFO org.apache.hadoop.ipc.Server: IPC Server >> handler 9 on 8020: starting >> 2009-10-14 11:09:53,444 INFO org.apache.hadoop.ipc.Server: IPC Server >> handler 5 on 8020: starting >> 2009-10-14 11:09:54,232 INFO org.apache.hadoop.ipc.Server: Error >> register getProtocolVersion >> java.lang.IllegalArgumentException: Duplicate >> metricsName:getProtocolVersion >> at >> org.apache.hadoop.metrics.util.MetricsRegistry.add(MetricsRegistry.java:53) >> at >> org.apache.hadoop.metrics.util.MetricsTimeVaryingRate.<init>(MetricsTimeVaryingRate.java:89) >> at >> org.apache.hadoop.metrics.util.MetricsTimeVaryingRate.<init>(MetricsTimeVaryingRate.java:99) >> at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:523) >> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:959) >> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:955) >> at java.security.AccessController.doPrivileged(Native Method) >> at javax.security.auth.Subject.doAs(Subject.java:396) >> at org.apache.hadoop.ipc.Server$Handler.run(Server.java:953) >> 2009-10-14 11:09:54,667 INFO org.apache.hadoop.hdfs.StateChange: >> BLOCK* NameSystem.registerDatanode: node registration from >> 192.168.76.3:50010 storage >> DS-551038728-192.168.76.3-50010-1255511394664 >> 2009-10-14 11:09:54,668 INFO org.apache.hadoop.net.NetworkTopology: >> Adding a new node: /default-rack/192.168.76.3:50010 >> 2009-10-14 11:09:54,673 INFO org.apache.hadoop.hdfs.StateChange: >> BLOCK* NameSystem.registerDatanode: node registration from >> 192.168.76.2:50010 storage >> DS-681115827-192.168.76.2-50010-1255511394673 >> 2009-10-14 11:09:54,674 INFO org.apache.hadoop.net.NetworkTopology: >> Adding a new node: /default-rack/192.168.76.2:50010 >> 2009-10-14 11:09:54,689 INFO org.apache.hadoop.hdfs.StateChange: >> BLOCK* NameSystem.registerDatanode: node registration from >> 192.168.76.7:50010 storage >> DS-1731752631-192.168.76.7-50010-1255511394688 >> 2009-10-14 11:09:54,689 INFO org.apache.hadoop.net.NetworkTopology: >> Adding a new node: /default-rack/192.168.76.7:50010 >> 2009-10-14 11:09:54,701 INFO org.apache.hadoop.hdfs.StateChange: >> BLOCK* NameSystem.registerDatanode: node registration from >> 192.168.76.6:50010 storage >> DS-338163265-192.168.76.6-50010-1255511394698 >> 2009-10-14 11:09:54,701 INFO org.apache.hadoop.net.NetworkTopology: >> Adding a new node: /default-rack/192.168.76.6:50010 >> 2009-10-14 11:09:59,688 INFO org.apache.hadoop.hdfs.StateChange: >> BLOCK* NameSystem.registerDatanode: node registration from >> 192.168.76.1:50010 storage >> DS-988595128-192.168.76.1-50010-1255511399688 >> 2009-10-14 11:09:59,689 INFO org.apache.hadoop.net.NetworkTopology: >> Adding a new node: /default-rack/192.168.76.1:50010 >> 2009-10-14 11:10:04,277 INFO org.apache.hadoop.hdfs.StateChange: >> BLOCK* NameSystem.registerDatanode: node registration from >> 192.168.76.9:50010 storage >> DS-1096719686-192.168.76.9-50010-1255511404276 >> 2009-10-14 11:10:04,278 INFO org.apache.hadoop.net.NetworkTopology: >> Adding a new node: /default-rack/192.168.76.9:50010 >> 2009-10-14 11:10:04,289 INFO org.apache.hadoop.hdfs.StateChange: >> BLOCK* NameSystem.registerDatanode: node registration from >> 192.168.76.5:50010 storage >> DS-1405051412-192.168.76.5-50010-1255511404288 >> 2009-10-14 11:10:04,289 INFO org.apache.hadoop.net.NetworkTopology: >> Adding a new node: /default-rack/192.168.76.5:50010 >> 2009-10-14 11:10:05,961 INFO org.apache.hadoop.hdfs.StateChange: >> BLOCK* NameSystem.registerDatanode: node registration from >> 192.168.76.8:50010 storage >> DS-1138617755-192.168.76.8-50010-1255511405960 >> 2009-10-14 11:10:05,961 INFO org.apache.hadoop.net.NetworkTopology: >> Adding a new node: /default-rack/192.168.76.8:50010 >> 2009-10-14 11:10:06,154 INFO org.apache.hadoop.hdfs.StateChange: >> BLOCK* NameSystem.registerDatanode: node registration from >> 192.168.76.4:50010 storage >> DS-569548170-192.168.76.4-50010-1255511406153 >> 2009-10-14 11:10:06,154 INFO org.apache.hadoop.net.NetworkTopology: >> Adding a new node: /default-rack/192.168.76.4:50010 >> 2009-10-14 11:10:14,480 INFO org.apache.hadoop.ipc.Client: Retrying >> connect to server: cluma.gbif.clu/192.168.76.254:8020. Already tried 0 >> time(s). >> 2009-10-14 11:10:35,483 INFO org.apache.hadoop.ipc.Client: Retrying >> connect to server: cluma.gbif.clu/192.168.76.254:8020. Already tried 1 >> time(s). >> 2009-10-14 11:10:56,486 INFO org.apache.hadoop.ipc.Client: Retrying >> connect to server: cluma.gbif.clu/192.168.76.254:8020. Already tried 2 >> time(s). >> 2009-10-14 11:11:17,489 INFO org.apache.hadoop.ipc.Client: Retrying >> connect to server: cluma.gbif.clu/192.168.76.254:8020. Already tried 3 >> time(s). >> 2009-10-14 11:11:38,492 INFO org.apache.hadoop.ipc.Client: Retrying >> connect to server: cluma.gbif.clu/192.168.76.254:8020. Already tried 4 >> time(s). >> 2009-10-14 11:11:59,495 INFO org.apache.hadoop.ipc.Client: Retrying >> connect to server: cluma.gbif.clu/192.168.76.254:8020. Already tried 5 >> time(s). >> 2009-10-14 11:12:20,498 INFO org.apache.hadoop.ipc.Client: Retrying >> connect to server: cluma.gbif.clu/192.168.76.254:8020. Already tried 6 >> time(s). >> 2009-10-14 11:12:41,501 INFO org.apache.hadoop.ipc.Client: Retrying >> connect to server: cluma.gbif.clu/192.168.76.254:8020. Already tried 7 >> time(s). >> 2009-10-14 11:13:02,504 INFO org.apache.hadoop.ipc.Client: Retrying >> connect to server: cluma.gbif.clu/192.168.76.254:8020. Already tried 8 >> time(s). >> 2009-10-14 11:13:23,506 INFO org.apache.hadoop.ipc.Client: Retrying >> connect to server: cluma.gbif.clu/192.168.76.254:8020. Already tried 9 >> time(s). >> etc. > >