I configured dfs.http.address on SNN's hdfs-site.xml but still gets: /************************************************************ STARTUP_MSG: Starting SecondaryNameNode STARTUP_MSG: host = hadoop01/192.168.0.11 STARTUP_MSG: args = [-checkpoint, force] STARTUP_MSG: version = 1.0.3 STARTUP_MSG: build = https://svn.apache.org/repos/asf/hadoop/common/branches/branch-1.0 -r 1335192; compiled by 'hortonfo' on Tue May 8 20:31:25 UTC 2012 ************************************************************/ 12/06/04 13:34:24 INFO namenode.SecondaryNameNode: Starting web server as: hadoop 12/06/04 13:34:24 INFO mortbay.log: Logging to org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via org.mortbay.log.Slf4jLog 12/06/04 13:34:24 INFO http.HttpServer: Added global filtersafety (class=org.apache.hadoop.http.HttpServer$QuotingInputFilter) 12/06/04 13:34:24 INFO http.HttpServer: Port returned by webServer.getConnectors()[0].getLocalPort() before open() is -1. Opening the listener on 50090 12/06/04 13:34:24 INFO http.HttpServer: listener.getLocalPort() returned 50090 webServer.getConnectors()[0].getLocalPort() returned 50090 12/06/04 13:34:24 INFO http.HttpServer: Jetty bound to port 50090 12/06/04 13:34:24 INFO mortbay.log: jetty-6.1.26 12/06/04 13:34:25 INFO mortbay.log: Started SelectChannelConnector@0.0.0.0:50090 12/06/04 13:34:25 INFO namenode.SecondaryNameNode: Web server init done 12/06/04 13:34:25 INFO namenode.SecondaryNameNode: Secondary Web-server up at: 0.0.0.0:50090 12/06/04 13:34:25 INFO namenode.SecondaryNameNode: Secondary image servlet up at: 0.0.0.0:50090 12/06/04 13:34:25 WARN namenode.SecondaryNameNode: Checkpoint Period :3600 secs (60 min) 12/06/04 13:34:25 WARN namenode.SecondaryNameNode: Log Size Trigger :67108864 bytes (65536 KB) 12/06/04 13:34:25 ERROR security.UserGroupInformation: PriviledgedActionException as:hadoop cause:java.net.ConnectException: Connection refused 12/06/04 13:34:25 ERROR namenode.SecondaryNameNode: checkpoint: Connection refused 12/06/04 13:34:25 INFO namenode.SecondaryNameNode: SHUTDOWN_MSG: /************************************************************ SHUTDOWN_MSG: Shutting down SecondaryNameNode at hadoop01/192.168.0.11 ************************************************************/
-----Original Message----- From: praveenesh kumar [mailto:praveen...@gmail.com] Sent: lunes, 04 de junio de 2012 13:15 To: common-user@hadoop.apache.org Subject: Re: SecondaryNameNode not connecting to NameNode : PriviledgedActionException I am not sure what could be the exact issue but when configuring secondary NN to NN, you need to tell your SNN where the actual NN resides. Try adding - dfs.http.address on your secondary namenode machine having value as <NN:port> on hdfs-site.xml Port should be on which your NN url is opening - means your NN web browser http port. Regards, Praveenesh On Mon, Jun 4, 2012 at 4:37 PM, <ramon....@accenture.com> wrote: > Hello. I'm facing a issue when trying to configure my > SecondaryNameNode on a different machine than my NameNode. When both > are on the same machine everything works fine but after moving the secondary > to a new machine I get: > > 2012-05-28 09:57:36,832 ERROR > org.apache.hadoop.security.UserGroupInformation: > PriviledgedActionException as:hadoop cause:java.net.ConnectException: > Connection refused > 2012-05-28 09:57:36,832 ERROR > org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode: Exception in > doCheckpoint: > 2012-05-28 09:57:36,834 ERROR > org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode: > java.net.ConnectException: Connection refused > at java.net.PlainSocketImpl.socketConnect(Native Method) > at > java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:327) > at > java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:191) > at > java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:180) > at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:384) > at java.net.Socket.connect(Socket.java:546) > at java.net.Socket.connect(Socket.java:495) > at sun.net.NetworkClient.doConnect(NetworkClient.java:178) > at sun.net.www.http.HttpClient.openServer(HttpClient.java:409) > at sun.net.www.http.HttpClient.openServer(HttpClient.java:530) > at sun.net.www.http.HttpClient.<init>(HttpClient.java:240) > at sun.net.www.http.HttpClient.New(HttpClient.java:321) > at sun.net.www.http.HttpClient.New(HttpClient.java:338) > at > sun.net.www.protocol.http.HttpURLConnection.getNewHttpClient(HttpURLConnection.java:935) > at > sun.net.www.protocol.http.HttpURLConnection.plainConnect(HttpURLConnection.java:876) > at > sun.net.www.protocol.http.HttpURLConnection.connect(HttpURLConnection. > java:801) > > Is there any configuration I'm missing? At this point my > mapred-site.xml is very simple just: > > <?xml version="1.0"?> > <?xml-stylesheet type="text/xsl" href="configuration.xsl"?> > <configuration> <property> > <name>mapred.job.tracker</name> > <value>hadoop00:9001</value> > </property> > <property> > <name>mapred.system.dir</name> > <value>/home/hadoop/mapred/system</value> > </property> > <property> > <name>mapred.local.dir</name> > <value>/home/hadoop/mapred/local</value> > </property> > <property> > <name>mapred.jobtracker.taskScheduler</name> > <value>org.apache.hadoop.mapred.FairScheduler</value> > </property> > <property> > <name>mapred.fairscheduler.allocation.file</name> > <value>/home/hadoop/hadoop/conf/fairscheduler.xml</value> > </property> > </configuration> > > > > ________________________________ > Subject to local law, communications with Accenture and its affiliates > including telephone calls and emails (including content), may be > monitored by our systems for the purposes of security and the > assessment of internal compliance with Accenture policy. > > ______________________________________________________________________ > ________________ > > www.accenture.com > ________________________________ Subject to local law, communications with Accenture and its affiliates including telephone calls and emails (including content), may be monitored by our systems for the purposes of security and the assessment of internal compliance with Accenture policy. ______________________________________________________________________________________ www.accenture.com