see the error .. it says not formatted did you press Y or y ? try again :)
On Tue, Dec 24, 2013 at 8:35 PM, Sitaraman Vilayannur < vrsitaramanietfli...@gmail.com> wrote: > Hi Nitin, > Even after formatting using hdfs namenode -format, i keep seeing namenode > not formatted in the logs when i try to start namenode........ > 12/24 20:33:26 INFO namenode.FSNamesystem: supergroup=supergroup > 13/12/24 20:33:26 INFO namenode.FSNamesystem: isPermissionEnabled=true > 13/12/24 20:33:26 INFO namenode.NameNode: Caching file names occuring more > than 10 times > 13/12/24 20:33:26 INFO namenode.NNStorage: Storage directory > /usr/local/Software/hadoop-0.23.10/data/hdfs/namenode has been successfully > formatted. > 13/12/24 20:33:26 INFO namenode.FSImage: Saving image file > /usr/local/Software/hadoop-0.23.10/data/hdfs/namenode/current/fsimage.ckpt_0000000000000000000 > using no compression > 13/12/24 20:33:26 INFO namenode.FSImage: Image file of size 124 saved in 0 > seconds. > 13/12/24 20:33:26 INFO namenode.NNStorageRetentionManager: Going to retain > 1 images with txid >= 0 > 13/12/24 20:33:26 INFO util.ExitUtil: Exiting with status 0 > 13/12/24 20:33:26 INFO namenode.NameNode: SHUTDOWN_MSG: > > /************************************************************ > SHUTDOWN_MSG: Shutting down NameNode at localhost.localdomain/127.0.0.1 > ************************************************************/ > > > 2013-12-24 20:33:46,337 INFO org.apache.hadoop.hdfs.server.common.Storage: > Lock on /usr/local/Software/hadoop-2.2.0/data/hdfs/namenode/in_use.lock > acquired by nodename 7518@localhost.localdomain > 2013-12-24 20:33:46,339 INFO org.mortbay.log: Stopped > SelectChannelConnector@0.0.0.0:50070 > 2013-12-24 20:33:46,340 INFO > org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Stopping NameNode > metrics system... > 2013-12-24 20:33:46,340 INFO > org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics system > stopped. > 2013-12-24 20:33:46,340 INFO > org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics system > shutdown complete. > 2013-12-24 20:33:46,340 FATAL > org.apache.hadoop.hdfs.server.namenode.NameNode: Exception in namenode join > java.io.IOException: NameNode is not formatted. > at > org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:210) > > at > org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFSImage(FSNamesystem.java:787) > at > org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFromDisk(FSNamesystem.java:568) > at > org.apache.hadoop.hdfs.server.namenode.NameNode.loadNamesystem(NameNode.java:443) > at > org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:491) > at > org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:684) > at > org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:669) > at > org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1254) > at > org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1320) > 2013-12-24 20:33:46,342 INFO org.apache.hadoop.util.ExitUtil: Exiting with > status 1 > 2013-12-24 20:33:46,343 INFO > org.apache.hadoop.hdfs.server.namenode.NameNode: SHUTDOWN_MSG: > > /************************************************************ > SHUTDOWN_MSG: Shutting down NameNode at localhost.localdomain/127.0.0.1 > ************************************************************/ > > > > > On Tue, Dec 24, 2013 at 3:13 PM, Nitin Pawar <nitinpawar...@gmail.com>wrote: > >> the issue here is you tried one version of hadoop and then changed to a >> different version. >> >> You can not do that directly with hadoop. You need to follow a process >> while upgrading hadoop versions. >> >> For now as you are just starting with hadoop, I would recommend just run >> a dfs format and start the hdfs again >> >> >> On Tue, Dec 24, 2013 at 2:57 PM, Sitaraman Vilayannur < >> vrsitaramanietfli...@gmail.com> wrote: >> >>> When i run namenode with upgrade option i get the following error and >>> and namenode dosent start... >>> 2013-12-24 14:48:38,595 INFO org.apache.hadoop.hdfs.StateChange: >>> STATE* Network topology has 0 racks and 0 datanodes >>> 2013-12-24 14:48:38,595 INFO org.apache.hadoop.hdfs.StateChange: >>> STATE* UnderReplicatedBlocks has 0 blocks >>> 2013-12-24 14:48:38,631 INFO org.apache.hadoop.ipc.Server: IPC Server >>> Responder: starting >>> 2013-12-24 14:48:38,632 INFO org.apache.hadoop.ipc.Server: IPC Server >>> listener on 9000: starting >>> 2013-12-24 14:48:38,633 INFO >>> org.apache.hadoop.hdfs.server.namenode.NameNode: NameNode RPC up at: >>> 192.168.1.2/192.168.1.2:9000 >>> 2013-12-24 14:48:38,633 INFO >>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Starting services >>> required for active state >>> 2013-12-24 14:50:50,060 ERROR >>> org.apache.hadoop.hdfs.server.namenode.NameNode: RECEIVED SIGNAL 15: >>> SIGTERM >>> 2013-12-24 14:50:50,062 INFO >>> org.apache.hadoop.hdfs.server.namenode.NameNode: SHUTDOWN_MSG: >>> /************************************************************ >>> SHUTDOWN_MSG: Shutting down NameNode at localhost.localdomain/127.0.0.1 >>> ************************************************************/ >>> >>> >>> On 12/24/13, Sitaraman Vilayannur <vrsitaramanietfli...@gmail.com> >>> wrote: >>> > Found it, >>> > I get the following error on starting namenode in 2.2 >>> > >>> 10/contrib/capacity-scheduler/*.jar:/usr/local/Software/hadoop-0.23.10/contrib/capacity-scheduler/*.jar:/usr/local/Software/hadoop-0.23.10/contrib/capacity-scheduler/*.jar >>> > STARTUP_MSG: build = https://svn.apache.org/repos/asf/hadoop/common >>> > -r 1529768; compiled by 'hortonmu' on 2013-10-07T06:28Z >>> > STARTUP_MSG: java = 1.7.0_45 >>> > ************************************************************/ >>> > 2013-12-24 13:25:48,876 INFO >>> > org.apache.hadoop.hdfs.server.namenode.NameNode: registered UNIX >>> > signal handlers for [TERM, HUP, INT] >>> > 2013-12-24 13:25:49,042 INFO >>> > org.apache.hadoop.metrics2.impl.MetricsConfig: loaded properties from >>> > hadoop-metrics2.properties >>> > 2013-12-24 13:25:49,102 INFO >>> > org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Scheduled snapshot >>> > period at 10 second(s). >>> > 2013-12-24 13:25:49,102 INFO >>> > org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics >>> > system started >>> > 2013-12-24 13:25:49,232 WARN org.apache.hadoop.util.NativeCodeLoader: >>> > Unable to load native-hadoop library for your platform... using >>> > builtin-java classes where applicable >>> > 2013-12-24 13:25:49,375 INFO org.mortbay.log: Logging to >>> > org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via >>> > org.mortbay.log.Slf4jLog >>> > 2013-12-24 13:25:49,410 INFO org.apache.hadoop.http.HttpServer: Added >>> > global filter 'safety' >>> > (class=org.apache.hadoop.http.HttpServer$QuotingInputFilter) >>> > 2013-12-24 13:25:49,412 INFO org.apache.hadoop.http.HttpServer: Added >>> > filter static_user_filter >>> > (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) >>> > to context hdfs >>> > 2013-12-24 13:25:49,412 INFO org.apache.hadoop.http.HttpServer: Added >>> > filter static_user_filter >>> > (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) >>> > to context static >>> > 2013-12-24 13:25:49,412 INFO org.apache.hadoop.http.HttpServer: Added >>> > filter static_user_filter >>> > (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) >>> > to context logs >>> > 2013-12-24 13:25:49,422 INFO org.apache.hadoop.http.HttpServer: >>> > dfs.webhdfs.enabled = false >>> > 2013-12-24 13:25:49,432 INFO org.apache.hadoop.http.HttpServer: Jetty >>> > bound to port 50070 >>> > 2013-12-24 13:25:49,432 INFO org.mortbay.log: jetty-6.1.26 >>> > 2013-12-24 13:25:49,459 WARN org.mortbay.log: Can't reuse >>> > /tmp/Jetty_0_0_0_0_50070_hdfs____w2cu08, using >>> > /tmp/Jetty_0_0_0_0_50070_hdfs____w2cu08_2787234685293301311 >>> > 2013-12-24 13:25:49,610 INFO org.mortbay.log: Started >>> > SelectChannelConnector@0.0.0.0:50070 >>> > 2013-12-24 13:25:49,611 INFO >>> > org.apache.hadoop.hdfs.server.namenode.NameNode: Web-server up at: >>> > 0.0.0.0:50070 >>> > 2013-12-24 13:25:49,628 WARN >>> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Only one image >>> > storage directory (dfs.namenode.name.dir) configured. Beware of >>> > dataloss due to lack of redundant storage directories! >>> > 2013-12-24 13:25:49,628 WARN >>> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Only one >>> > namespace edits storage directory (dfs.namenode.edits.dir) configured. >>> > Beware of dataloss due to lack of redundant storage directories! >>> > 2013-12-24 13:25:49,668 INFO >>> > org.apache.hadoop.hdfs.server.namenode.HostFileManager: read includes: >>> > HostSet( >>> > ) >>> > 2013-12-24 13:25:49,669 INFO >>> > org.apache.hadoop.hdfs.server.namenode.HostFileManager: read excludes: >>> > HostSet( >>> > ) >>> > 2013-12-24 13:25:49,670 INFO >>> > org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: >>> > dfs.block.invalidate.limit=1000 >>> > 2013-12-24 13:25:49,672 INFO org.apache.hadoop.util.GSet: Computing >>> > capacity for map BlocksMap >>> > 2013-12-24 13:25:49,672 INFO org.apache.hadoop.util.GSet: VM type >>> = >>> > 64-bit >>> > 2013-12-24 13:25:49,673 INFO org.apache.hadoop.util.GSet: 2.0% max >>> > memory = 889 MB >>> > 2013-12-24 13:25:49,673 INFO org.apache.hadoop.util.GSet: capacity >>> > = 2^21 = 2097152 entries >>> > 2013-12-24 13:25:49,677 INFO >>> > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: >>> > dfs.block.access.token.enable=false >>> > 2013-12-24 13:25:49,677 INFO >>> > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: >>> > defaultReplication = 1 >>> > 2013-12-24 13:25:49,677 INFO >>> > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: >>> > maxReplication = 512 >>> > 2013-12-24 13:25:49,677 INFO >>> > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: >>> > minReplication = 1 >>> > 2013-12-24 13:25:49,677 INFO >>> > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: >>> > maxReplicationStreams = 2 >>> > 2013-12-24 13:25:49,677 INFO >>> > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: >>> > shouldCheckForEnoughRacks = false >>> > 2013-12-24 13:25:49,677 INFO >>> > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: >>> > replicationRecheckInterval = 3000 >>> > 2013-12-24 13:25:49,677 INFO >>> > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: >>> > encryptDataTransfer = false >>> > 2013-12-24 13:25:49,681 INFO >>> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: fsOwner >>> > = sitaraman (auth:SIMPLE) >>> > 2013-12-24 13:25:49,681 INFO >>> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: supergroup >>> > = supergroup >>> > 2013-12-24 13:25:49,681 INFO >>> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: >>> > isPermissionEnabled = true >>> > 2013-12-24 13:25:49,681 INFO >>> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: HA Enabled: false >>> > 2013-12-24 13:25:49,682 INFO >>> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Append Enabled: >>> > true >>> > 2013-12-24 13:25:49,801 INFO org.apache.hadoop.util.GSet: Computing >>> > capacity for map INodeMap >>> > 2013-12-24 13:25:49,801 INFO org.apache.hadoop.util.GSet: VM type >>> = >>> > 64-bit >>> > 2013-12-24 13:25:49,801 INFO org.apache.hadoop.util.GSet: 1.0% max >>> > memory = 889 MB >>> > 2013-12-24 13:25:49,801 INFO org.apache.hadoop.util.GSet: capacity >>> > = 2^20 = 1048576 entries >>> > 2013-12-24 13:25:49,802 INFO >>> > org.apache.hadoop.hdfs.server.namenode.NameNode: Caching file names >>> > occuring more than 10 times >>> > 2013-12-24 13:25:49,804 INFO >>> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: >>> > dfs.namenode.safemode.threshold-pct = 0.9990000128746033 >>> > 2013-12-24 13:25:49,804 INFO >>> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: >>> > dfs.namenode.safemode.min.datanodes = 0 >>> > 2013-12-24 13:25:49,804 INFO >>> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: >>> > dfs.namenode.safemode.extension = 30000 >>> > 2013-12-24 13:25:49,805 INFO >>> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Retry cache on >>> > namenode is enabled >>> > 2013-12-24 13:25:49,805 INFO >>> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Retry cache will >>> > use 0.03 of total heap and retry cache entry expiry time is 600000 >>> > millis >>> > 2013-12-24 13:25:49,807 INFO org.apache.hadoop.util.GSet: Computing >>> > capacity for map Namenode Retry Cache >>> > 2013-12-24 13:25:49,807 INFO org.apache.hadoop.util.GSet: VM type >>> = >>> > 64-bit >>> > 2013-12-24 13:25:49,807 INFO org.apache.hadoop.util.GSet: >>> > 0.029999999329447746% max memory = 889 MB >>> > 2013-12-24 13:25:49,807 INFO org.apache.hadoop.util.GSet: capacity >>> > = 2^15 = 32768 entries >>> > 2013-12-24 13:25:49,816 INFO >>> > org.apache.hadoop.hdfs.server.common.Storage: Lock on >>> > /usr/local/Software/hadoop-0.23.10/data/hdfs/namenode/in_use.lock >>> > acquired by nodename 19170@localhost.localdomain >>> > 2013-12-24 13:25:49,861 INFO org.mortbay.log: Stopped >>> > SelectChannelConnector@0.0.0.0:50070 >>> > 2013-12-24 13:25:49,964 INFO >>> > org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Stopping NameNode >>> > metrics system... >>> > 2013-12-24 13:25:49,965 INFO >>> > org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics >>> > system stopped. >>> > 2013-12-24 13:25:49,965 INFO >>> > org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics >>> > system shutdown complete. >>> > 2013-12-24 13:25:49,965 FATAL >>> > org.apache.hadoop.hdfs.server.namenode.NameNode: Exception in namenode >>> > join >>> > java.io.IOException: >>> > File system image contains an old layout version -39. >>> > An upgrade to version -47 is required. >>> > Please restart NameNode with -upgrade option. >>> > at >>> > >>> org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:221) >>> > at >>> > >>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFSImage(FSNamesystem.java:787) >>> > at >>> > >>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFromDisk(FSNamesystem.java:568) >>> > at >>> > >>> org.apache.hadoop.hdfs.server.namenode.NameNode.loadNamesystem(NameNode.java:443) >>> > at >>> > >>> org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:491) >>> > at >>> > >>> org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:684) >>> > at >>> > >>> org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:669) >>> > at >>> > >>> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1254) >>> > at >>> > >>> org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1320) >>> > 2013-12-24 13:25:49,967 INFO org.apache.hadoop.util.ExitUtil: Exiting >>> > with status 1 >>> > 2013-12-24 13:25:49,968 INFO >>> > org.apache.hadoop.hdfs.server.namenode.NameNode: SHUTDOWN_MSG: >>> > /************************************************************ >>> > SHUTDOWN_MSG: Shutting down NameNode at localhost.localdomain/ >>> 127.0.0.1 >>> > ************************************************************/ >>> > >>> > On 12/24/13, Sitaraman Vilayannur <vrsitaramanietfli...@gmail.com> >>> wrote: >>> >> The line beginning with ulimit that i have appended below, i thought >>> >> was the log file? >>> >> >>> /usr/local/Software/hadoop-2.2.0/logs/hadoop-sitaraman-namenode-localhost.localdomain.out >>> >> Sitaraman >>> >> On 12/24/13, Nitin Pawar <nitinpawar...@gmail.com> wrote: >>> >>> Without log, very hard to guess what's happening. >>> >>> >>> >>> Can you clean up the log directory and then start over and check for >>> the >>> >>> logs again. >>> >>> >>> >>> >>> >>> On Tue, Dec 24, 2013 at 11:44 AM, Sitaraman Vilayannur < >>> >>> vrsitaramanietfli...@gmail.com> wrote: >>> >>> >>> >>>> Hi Nitin, >>> >>>> I moved to the release 2.2.0 on starting node manager it remains >>> >>>> silent without errors but nodemanager dosent start....while it does >>> in >>> >>>> the earlier 0.23 version >>> >>>> >>> >>>> >>> >>>> ./hadoop-daemon.sh start namenode >>> >>>> starting namenode, logging to >>> >>>> >>> >>>> >>> /usr/local/Software/hadoop-2.2.0/logs/hadoop-sitaraman-namenode-localhost.localdomain.out >>> >>>> Java HotSpot(TM) 64-Bit Server VM warning: You have loaded library >>> >>>> /usr/local/Software/hadoop-2.2.0/lib/native/libhadoop.so.1.0.0 which >>> >>>> might have disabled stack guard. The VM will try to fix the stack >>> >>>> guard now. >>> >>>> It's highly recommended that you fix the library with 'execstack -c >>> >>>> <libfile>', or link it with '-z noexecstack'. >>> >>>> [sitaraman@localhost sbin]$ jps >>> >>>> 13444 Jps >>> >>>> [sitaraman@localhost sbin]$ vi >>> >>>> >>> >>>> >>> /usr/local/Software/hadoop-2.2.0/logs/hadoop-sitaraman-namenode-localhost.localdomain.out >>> >>>> >>> >>>> >>> >>>> ulimit -a for user sitaraman >>> >>>> core file size (blocks, -c) 0 >>> >>>> data seg size (kbytes, -d) unlimited >>> >>>> scheduling priority (-e) 0 >>> >>>> file size (blocks, -f) unlimited >>> >>>> pending signals (-i) 135104 >>> >>>> max locked memory (kbytes, -l) 32 >>> >>>> max memory size (kbytes, -m) unlimited >>> >>>> open files (-n) 1024 >>> >>>> pipe size (512 bytes, -p) 8 >>> >>>> POSIX message queues (bytes, -q) 819200 >>> >>>> real-time priority (-r) 0 >>> >>>> stack size (kbytes, -s) 10240 >>> >>>> cpu time (seconds, -t) unlimited >>> >>>> max user processes (-u) 135104 >>> >>>> virtual memory (kbytes, -v) unlimited >>> >>>> file locks (-x) unlimited >>> >>>> >>> >>>> >>> >>>> On 12/24/13, Nitin Pawar <nitinpawar...@gmail.com> wrote: >>> >>>> > For now you can ignore this warning, >>> >>>> > it was your first program so you can try building other things and >>> >>>> > slowly >>> >>>> > run the commands mentioned the log message to fix these small >>> >>>> > warnings. >>> >>>> > >>> >>>> > >>> >>>> > On Tue, Dec 24, 2013 at 10:07 AM, Sitaraman Vilayannur < >>> >>>> > vrsitaramanietfli...@gmail.com> wrote: >>> >>>> > >>> >>>> >> Thanks Nitin, That worked, >>> >>>> >> When i run the Pi example, i get the following warning at the >>> end, >>> >>>> >> what must i do about this warning....thanks much for your help. >>> >>>> >> Sitaraman >>> >>>> >> inished in 20.82 seconds >>> >>>> >> Java HotSpot(TM) 64-Bit Server VM warning: You have loaded >>> library >>> >>>> >> /usr/local/Software/hadoop-0.23.10/lib/native/libhadoop.so.1.0.0 >>> >>>> >> which >>> >>>> >> might have disabled stack guard. The VM will try to fix the stack >>> >>>> >> guard now. >>> >>>> >> It's highly recommended that you fix the library with 'execstack >>> -c >>> >>>> >> <libfile>', or link it with '-z noexecstack'. >>> >>>> >> 13/12/24 10:05:19 WARN util.NativeCodeLoader: Unable to load >>> >>>> >> native-hadoop library for your platform... using builtin-java >>> >>>> >> classes >>> >>>> >> where applicable >>> >>>> >> Estimated value of Pi is 3.14127500000000000000 >>> >>>> >> [sitaraman@localhost mapreduce]$ >>> >>>> >> >>> >>>> >> On 12/23/13, Nitin Pawar <nitinpawar...@gmail.com> wrote: >>> >>>> >> > Can you try starting the process as non root user. >>> >>>> >> > Give proper permissions to the user and start it as a different >>> >>>> >> > user. >>> >>>> >> > >>> >>>> >> > Thanks, >>> >>>> >> > Nitin >>> >>>> >> > >>> >>>> >> > >>> >>>> >> > On Mon, Dec 23, 2013 at 2:15 PM, Sitaraman Vilayannur < >>> >>>> >> > vrsitaramanietfli...@gmail.com> wrote: >>> >>>> >> > >>> >>>> >> >> Hi, >>> >>>> >> >> When i attempt to start nodemanager i get the following >>> error. >>> >>>> >> >> Any >>> >>>> >> >> help >>> >>>> >> >> appreciated. I was able to start resource manager datanode, >>> >>>> namenode >>> >>>> >> >> and >>> >>>> >> >> secondarynamenode, >>> >>>> >> >> >>> >>>> >> >> >>> >>>> >> >> ./yarn-daemon.sh start nodemanager >>> >>>> >> >> starting nodemanager, logging to >>> >>>> >> >> >>> >>>> >> >>> >>>> >>> /usr/local/Software/hadoop-0.23.10/logs/yarn-root-nodemanager-localhost.localdomain.out >>> >>>> >> >> Unrecognized option: -jvm >>> >>>> >> >> Error: Could not create the Java Virtual Machine. >>> >>>> >> >> Error: A fatal exception has occurred. Program will exit. >>> >>>> >> >> [root@localhost sbin]# emacs >>> >>>> >> >> >>> >>>> >> >>> >>>> >>> /usr/local/Software/hadoop-0.23.10/logs/yarn-root-nodemanager-localhost.localdomain.out >>> >>>> >> >> & >>> >>>> >> >> [4] 29004 >>> >>>> >> >> [root@localhost sbin]# jps >>> >>>> >> >> 28402 SecondaryNameNode >>> >>>> >> >> 30280 Jps >>> >>>> >> >> 28299 DataNode >>> >>>> >> >> 6729 Main >>> >>>> >> >> 26044 ResourceManager >>> >>>> >> >> 28197 NameNode >>> >>>> >> >> >>> >>>> >> > >>> >>>> >> > >>> >>>> >> > >>> >>>> >> > -- >>> >>>> >> > Nitin Pawar >>> >>>> >> > >>> >>>> >> >>> >>>> > >>> >>>> > >>> >>>> > >>> >>>> > -- >>> >>>> > Nitin Pawar >>> >>>> > >>> >>>> >>> >>> >>> >>> >>> >>> >>> >>> -- >>> >>> Nitin Pawar >>> >>> >>> >> >>> > >>> >> >> >> >> -- >> Nitin Pawar >> > > -- Nitin Pawar