see the error .. it says not formatted
did you press Y or y ?
try again :)

On Tue, Dec 24, 2013 at 8:35 PM, Sitaraman Vilayannur <
vrsitaramanietfli...@gmail.com> wrote:

> Hi Nitin,
>  Even after formatting using hdfs namenode -format, i keep seeing namenode
> not formatted in the logs when i try to start namenode........
> 12/24 20:33:26 INFO namenode.FSNamesystem: supergroup=supergroup
> 13/12/24 20:33:26 INFO namenode.FSNamesystem: isPermissionEnabled=true
> 13/12/24 20:33:26 INFO namenode.NameNode: Caching file names occuring more
> than 10 times
> 13/12/24 20:33:26 INFO namenode.NNStorage: Storage directory
> /usr/local/Software/hadoop-0.23.10/data/hdfs/namenode has been successfully
> formatted.
> 13/12/24 20:33:26 INFO namenode.FSImage: Saving image file
> /usr/local/Software/hadoop-0.23.10/data/hdfs/namenode/current/fsimage.ckpt_0000000000000000000
> using no compression
> 13/12/24 20:33:26 INFO namenode.FSImage: Image file of size 124 saved in 0
> seconds.
> 13/12/24 20:33:26 INFO namenode.NNStorageRetentionManager: Going to retain
> 1 images with txid >= 0
> 13/12/24 20:33:26 INFO util.ExitUtil: Exiting with status 0
> 13/12/24 20:33:26 INFO namenode.NameNode: SHUTDOWN_MSG:
>
> /************************************************************
> SHUTDOWN_MSG: Shutting down NameNode at localhost.localdomain/127.0.0.1
> ************************************************************/
>
>
> 2013-12-24 20:33:46,337 INFO org.apache.hadoop.hdfs.server.common.Storage:
> Lock on /usr/local/Software/hadoop-2.2.0/data/hdfs/namenode/in_use.lock
> acquired by nodename 7518@localhost.localdomain
> 2013-12-24 20:33:46,339 INFO org.mortbay.log: Stopped
> SelectChannelConnector@0.0.0.0:50070
> 2013-12-24 20:33:46,340 INFO
> org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Stopping NameNode
> metrics system...
> 2013-12-24 20:33:46,340 INFO
> org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics system
> stopped.
> 2013-12-24 20:33:46,340 INFO
> org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics system
> shutdown complete.
> 2013-12-24 20:33:46,340 FATAL
> org.apache.hadoop.hdfs.server.namenode.NameNode: Exception in namenode join
> java.io.IOException: NameNode is not formatted.
>         at
> org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:210)
>
>         at
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFSImage(FSNamesystem.java:787)
>         at
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFromDisk(FSNamesystem.java:568)
>         at
> org.apache.hadoop.hdfs.server.namenode.NameNode.loadNamesystem(NameNode.java:443)
>         at
> org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:491)
>         at
> org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:684)
>         at
> org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:669)
>         at
> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1254)
>         at
> org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1320)
> 2013-12-24 20:33:46,342 INFO org.apache.hadoop.util.ExitUtil: Exiting with
> status 1
> 2013-12-24 20:33:46,343 INFO
> org.apache.hadoop.hdfs.server.namenode.NameNode: SHUTDOWN_MSG:
>
> /************************************************************
> SHUTDOWN_MSG: Shutting down NameNode at localhost.localdomain/127.0.0.1
> ************************************************************/
>
>
>
>
> On Tue, Dec 24, 2013 at 3:13 PM, Nitin Pawar <nitinpawar...@gmail.com>wrote:
>
>> the issue here is you tried one version of hadoop and then changed to a
>> different version.
>>
>> You can not do that directly with hadoop. You need to follow a process
>> while upgrading hadoop versions.
>>
>> For now as you are just starting with hadoop, I would recommend just run
>> a dfs format and start the hdfs again
>>
>>
>> On Tue, Dec 24, 2013 at 2:57 PM, Sitaraman Vilayannur <
>> vrsitaramanietfli...@gmail.com> wrote:
>>
>>> When i run namenode with upgrade option i get the following error and
>>> and namenode dosent start...
>>> 2013-12-24 14:48:38,595 INFO org.apache.hadoop.hdfs.StateChange:
>>> STATE* Network topology has 0 racks and 0 datanodes
>>> 2013-12-24 14:48:38,595 INFO org.apache.hadoop.hdfs.StateChange:
>>> STATE* UnderReplicatedBlocks has 0 blocks
>>> 2013-12-24 14:48:38,631 INFO org.apache.hadoop.ipc.Server: IPC Server
>>> Responder: starting
>>> 2013-12-24 14:48:38,632 INFO org.apache.hadoop.ipc.Server: IPC Server
>>> listener on 9000: starting
>>> 2013-12-24 14:48:38,633 INFO
>>> org.apache.hadoop.hdfs.server.namenode.NameNode: NameNode RPC up at:
>>> 192.168.1.2/192.168.1.2:9000
>>> 2013-12-24 14:48:38,633 INFO
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Starting services
>>> required for active state
>>> 2013-12-24 14:50:50,060 ERROR
>>> org.apache.hadoop.hdfs.server.namenode.NameNode: RECEIVED SIGNAL 15:
>>> SIGTERM
>>> 2013-12-24 14:50:50,062 INFO
>>> org.apache.hadoop.hdfs.server.namenode.NameNode: SHUTDOWN_MSG:
>>> /************************************************************
>>> SHUTDOWN_MSG: Shutting down NameNode at localhost.localdomain/127.0.0.1
>>> ************************************************************/
>>>
>>>
>>> On 12/24/13, Sitaraman Vilayannur <vrsitaramanietfli...@gmail.com>
>>> wrote:
>>> > Found it,
>>> >  I get the following error on starting namenode in 2.2
>>> >
>>> 10/contrib/capacity-scheduler/*.jar:/usr/local/Software/hadoop-0.23.10/contrib/capacity-scheduler/*.jar:/usr/local/Software/hadoop-0.23.10/contrib/capacity-scheduler/*.jar
>>> > STARTUP_MSG:   build = https://svn.apache.org/repos/asf/hadoop/common
>>> > -r 1529768; compiled by 'hortonmu' on 2013-10-07T06:28Z
>>> > STARTUP_MSG:   java = 1.7.0_45
>>> > ************************************************************/
>>> > 2013-12-24 13:25:48,876 INFO
>>> > org.apache.hadoop.hdfs.server.namenode.NameNode: registered UNIX
>>> > signal handlers for [TERM, HUP, INT]
>>> > 2013-12-24 13:25:49,042 INFO
>>> > org.apache.hadoop.metrics2.impl.MetricsConfig: loaded properties from
>>> > hadoop-metrics2.properties
>>> > 2013-12-24 13:25:49,102 INFO
>>> > org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Scheduled snapshot
>>> > period at 10 second(s).
>>> > 2013-12-24 13:25:49,102 INFO
>>> > org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics
>>> > system started
>>> > 2013-12-24 13:25:49,232 WARN org.apache.hadoop.util.NativeCodeLoader:
>>> > Unable to load native-hadoop library for your platform... using
>>> > builtin-java classes where applicable
>>> > 2013-12-24 13:25:49,375 INFO org.mortbay.log: Logging to
>>> > org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via
>>> > org.mortbay.log.Slf4jLog
>>> > 2013-12-24 13:25:49,410 INFO org.apache.hadoop.http.HttpServer: Added
>>> > global filter 'safety'
>>> > (class=org.apache.hadoop.http.HttpServer$QuotingInputFilter)
>>> > 2013-12-24 13:25:49,412 INFO org.apache.hadoop.http.HttpServer: Added
>>> > filter static_user_filter
>>> > (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter)
>>> > to context hdfs
>>> > 2013-12-24 13:25:49,412 INFO org.apache.hadoop.http.HttpServer: Added
>>> > filter static_user_filter
>>> > (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter)
>>> > to context static
>>> > 2013-12-24 13:25:49,412 INFO org.apache.hadoop.http.HttpServer: Added
>>> > filter static_user_filter
>>> > (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter)
>>> > to context logs
>>> > 2013-12-24 13:25:49,422 INFO org.apache.hadoop.http.HttpServer:
>>> > dfs.webhdfs.enabled = false
>>> > 2013-12-24 13:25:49,432 INFO org.apache.hadoop.http.HttpServer: Jetty
>>> > bound to port 50070
>>> > 2013-12-24 13:25:49,432 INFO org.mortbay.log: jetty-6.1.26
>>> > 2013-12-24 13:25:49,459 WARN org.mortbay.log: Can't reuse
>>> > /tmp/Jetty_0_0_0_0_50070_hdfs____w2cu08, using
>>> > /tmp/Jetty_0_0_0_0_50070_hdfs____w2cu08_2787234685293301311
>>> > 2013-12-24 13:25:49,610 INFO org.mortbay.log: Started
>>> > SelectChannelConnector@0.0.0.0:50070
>>> > 2013-12-24 13:25:49,611 INFO
>>> > org.apache.hadoop.hdfs.server.namenode.NameNode: Web-server up at:
>>> > 0.0.0.0:50070
>>> > 2013-12-24 13:25:49,628 WARN
>>> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Only one image
>>> > storage directory (dfs.namenode.name.dir) configured. Beware of
>>> > dataloss due to lack of redundant storage directories!
>>> > 2013-12-24 13:25:49,628 WARN
>>> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Only one
>>> > namespace edits storage directory (dfs.namenode.edits.dir) configured.
>>> > Beware of dataloss due to lack of redundant storage directories!
>>> > 2013-12-24 13:25:49,668 INFO
>>> > org.apache.hadoop.hdfs.server.namenode.HostFileManager: read includes:
>>> > HostSet(
>>> > )
>>> > 2013-12-24 13:25:49,669 INFO
>>> > org.apache.hadoop.hdfs.server.namenode.HostFileManager: read excludes:
>>> > HostSet(
>>> > )
>>> > 2013-12-24 13:25:49,670 INFO
>>> > org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager:
>>> > dfs.block.invalidate.limit=1000
>>> > 2013-12-24 13:25:49,672 INFO org.apache.hadoop.util.GSet: Computing
>>> > capacity for map BlocksMap
>>> > 2013-12-24 13:25:49,672 INFO org.apache.hadoop.util.GSet: VM type
>>>   =
>>> > 64-bit
>>> > 2013-12-24 13:25:49,673 INFO org.apache.hadoop.util.GSet: 2.0% max
>>> > memory = 889 MB
>>> > 2013-12-24 13:25:49,673 INFO org.apache.hadoop.util.GSet: capacity
>>> >  = 2^21 = 2097152 entries
>>> > 2013-12-24 13:25:49,677 INFO
>>> > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager:
>>> > dfs.block.access.token.enable=false
>>> > 2013-12-24 13:25:49,677 INFO
>>> > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager:
>>> > defaultReplication         = 1
>>> > 2013-12-24 13:25:49,677 INFO
>>> > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager:
>>> > maxReplication             = 512
>>> > 2013-12-24 13:25:49,677 INFO
>>> > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager:
>>> > minReplication             = 1
>>> > 2013-12-24 13:25:49,677 INFO
>>> > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager:
>>> > maxReplicationStreams      = 2
>>> > 2013-12-24 13:25:49,677 INFO
>>> > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager:
>>> > shouldCheckForEnoughRacks  = false
>>> > 2013-12-24 13:25:49,677 INFO
>>> > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager:
>>> > replicationRecheckInterval = 3000
>>> > 2013-12-24 13:25:49,677 INFO
>>> > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager:
>>> > encryptDataTransfer        = false
>>> > 2013-12-24 13:25:49,681 INFO
>>> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: fsOwner
>>> >   = sitaraman (auth:SIMPLE)
>>> > 2013-12-24 13:25:49,681 INFO
>>> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: supergroup
>>> >   = supergroup
>>> > 2013-12-24 13:25:49,681 INFO
>>> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem:
>>> > isPermissionEnabled = true
>>> > 2013-12-24 13:25:49,681 INFO
>>> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: HA Enabled: false
>>> > 2013-12-24 13:25:49,682 INFO
>>> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Append Enabled:
>>> > true
>>> > 2013-12-24 13:25:49,801 INFO org.apache.hadoop.util.GSet: Computing
>>> > capacity for map INodeMap
>>> > 2013-12-24 13:25:49,801 INFO org.apache.hadoop.util.GSet: VM type
>>>   =
>>> > 64-bit
>>> > 2013-12-24 13:25:49,801 INFO org.apache.hadoop.util.GSet: 1.0% max
>>> > memory = 889 MB
>>> > 2013-12-24 13:25:49,801 INFO org.apache.hadoop.util.GSet: capacity
>>> >  = 2^20 = 1048576 entries
>>> > 2013-12-24 13:25:49,802 INFO
>>> > org.apache.hadoop.hdfs.server.namenode.NameNode: Caching file names
>>> > occuring more than 10 times
>>> > 2013-12-24 13:25:49,804 INFO
>>> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem:
>>> > dfs.namenode.safemode.threshold-pct = 0.9990000128746033
>>> > 2013-12-24 13:25:49,804 INFO
>>> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem:
>>> > dfs.namenode.safemode.min.datanodes = 0
>>> > 2013-12-24 13:25:49,804 INFO
>>> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem:
>>> > dfs.namenode.safemode.extension     = 30000
>>> > 2013-12-24 13:25:49,805 INFO
>>> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Retry cache on
>>> > namenode is enabled
>>> > 2013-12-24 13:25:49,805 INFO
>>> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Retry cache will
>>> > use 0.03 of total heap and retry cache entry expiry time is 600000
>>> > millis
>>> > 2013-12-24 13:25:49,807 INFO org.apache.hadoop.util.GSet: Computing
>>> > capacity for map Namenode Retry Cache
>>> > 2013-12-24 13:25:49,807 INFO org.apache.hadoop.util.GSet: VM type
>>>   =
>>> > 64-bit
>>> > 2013-12-24 13:25:49,807 INFO org.apache.hadoop.util.GSet:
>>> > 0.029999999329447746% max memory = 889 MB
>>> > 2013-12-24 13:25:49,807 INFO org.apache.hadoop.util.GSet: capacity
>>> >  = 2^15 = 32768 entries
>>> > 2013-12-24 13:25:49,816 INFO
>>> > org.apache.hadoop.hdfs.server.common.Storage: Lock on
>>> > /usr/local/Software/hadoop-0.23.10/data/hdfs/namenode/in_use.lock
>>> > acquired by nodename 19170@localhost.localdomain
>>> > 2013-12-24 13:25:49,861 INFO org.mortbay.log: Stopped
>>> > SelectChannelConnector@0.0.0.0:50070
>>> > 2013-12-24 13:25:49,964 INFO
>>> > org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Stopping NameNode
>>> > metrics system...
>>> > 2013-12-24 13:25:49,965 INFO
>>> > org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics
>>> > system stopped.
>>> > 2013-12-24 13:25:49,965 INFO
>>> > org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics
>>> > system shutdown complete.
>>> > 2013-12-24 13:25:49,965 FATAL
>>> > org.apache.hadoop.hdfs.server.namenode.NameNode: Exception in namenode
>>> > join
>>> > java.io.IOException:
>>> > File system image contains an old layout version -39.
>>> > An upgrade to version -47 is required.
>>> > Please restart NameNode with -upgrade option.
>>> >       at
>>> >
>>> org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:221)
>>> >       at
>>> >
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFSImage(FSNamesystem.java:787)
>>> >       at
>>> >
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFromDisk(FSNamesystem.java:568)
>>> >       at
>>> >
>>> org.apache.hadoop.hdfs.server.namenode.NameNode.loadNamesystem(NameNode.java:443)
>>> >       at
>>> >
>>> org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:491)
>>> >       at
>>> >
>>> org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:684)
>>> >       at
>>> >
>>> org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:669)
>>> >       at
>>> >
>>> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1254)
>>> >       at
>>> >
>>> org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1320)
>>> > 2013-12-24 13:25:49,967 INFO org.apache.hadoop.util.ExitUtil: Exiting
>>> > with status 1
>>> > 2013-12-24 13:25:49,968 INFO
>>> > org.apache.hadoop.hdfs.server.namenode.NameNode: SHUTDOWN_MSG:
>>> > /************************************************************
>>> > SHUTDOWN_MSG: Shutting down NameNode at localhost.localdomain/
>>> 127.0.0.1
>>> > ************************************************************/
>>> >
>>> > On 12/24/13, Sitaraman Vilayannur <vrsitaramanietfli...@gmail.com>
>>> wrote:
>>> >> The line beginning with ulimit that i have appended below, i thought
>>> >> was the log file?
>>> >>
>>> /usr/local/Software/hadoop-2.2.0/logs/hadoop-sitaraman-namenode-localhost.localdomain.out
>>> >> Sitaraman
>>> >> On 12/24/13, Nitin Pawar <nitinpawar...@gmail.com> wrote:
>>> >>> Without log, very hard to guess what's happening.
>>> >>>
>>> >>> Can you clean up the log directory and then start over and check for
>>> the
>>> >>> logs again.
>>> >>>
>>> >>>
>>> >>> On Tue, Dec 24, 2013 at 11:44 AM, Sitaraman Vilayannur <
>>> >>> vrsitaramanietfli...@gmail.com> wrote:
>>> >>>
>>> >>>> Hi Nitin,
>>> >>>>  I moved to the release 2.2.0 on starting node manager it remains
>>> >>>> silent without errors but nodemanager dosent start....while it does
>>> in
>>> >>>> the earlier 0.23 version
>>> >>>>
>>> >>>>
>>> >>>> ./hadoop-daemon.sh start namenode
>>> >>>> starting namenode, logging to
>>> >>>>
>>> >>>>
>>> /usr/local/Software/hadoop-2.2.0/logs/hadoop-sitaraman-namenode-localhost.localdomain.out
>>> >>>> Java HotSpot(TM) 64-Bit Server VM warning: You have loaded library
>>> >>>> /usr/local/Software/hadoop-2.2.0/lib/native/libhadoop.so.1.0.0 which
>>> >>>> might have disabled stack guard. The VM will try to fix the stack
>>> >>>> guard now.
>>> >>>> It's highly recommended that you fix the library with 'execstack -c
>>> >>>> <libfile>', or link it with '-z noexecstack'.
>>> >>>> [sitaraman@localhost sbin]$ jps
>>> >>>> 13444 Jps
>>> >>>> [sitaraman@localhost sbin]$ vi
>>> >>>>
>>> >>>>
>>> /usr/local/Software/hadoop-2.2.0/logs/hadoop-sitaraman-namenode-localhost.localdomain.out
>>> >>>>
>>> >>>>
>>> >>>> ulimit -a for user sitaraman
>>> >>>> core file size          (blocks, -c) 0
>>> >>>> data seg size           (kbytes, -d) unlimited
>>> >>>> scheduling priority             (-e) 0
>>> >>>> file size               (blocks, -f) unlimited
>>> >>>> pending signals                 (-i) 135104
>>> >>>> max locked memory       (kbytes, -l) 32
>>> >>>> max memory size         (kbytes, -m) unlimited
>>> >>>> open files                      (-n) 1024
>>> >>>> pipe size            (512 bytes, -p) 8
>>> >>>> POSIX message queues     (bytes, -q) 819200
>>> >>>> real-time priority              (-r) 0
>>> >>>> stack size              (kbytes, -s) 10240
>>> >>>> cpu time               (seconds, -t) unlimited
>>> >>>> max user processes              (-u) 135104
>>> >>>> virtual memory          (kbytes, -v) unlimited
>>> >>>> file locks                      (-x) unlimited
>>> >>>>
>>> >>>>
>>> >>>> On 12/24/13, Nitin Pawar <nitinpawar...@gmail.com> wrote:
>>> >>>> > For now you can ignore this warning,
>>> >>>> > it was your first program so you can try building other things and
>>> >>>> > slowly
>>> >>>> > run the commands mentioned the log message to fix these small
>>> >>>> > warnings.
>>> >>>> >
>>> >>>> >
>>> >>>> > On Tue, Dec 24, 2013 at 10:07 AM, Sitaraman Vilayannur <
>>> >>>> > vrsitaramanietfli...@gmail.com> wrote:
>>> >>>> >
>>> >>>> >> Thanks Nitin, That worked,
>>> >>>> >> When i run the Pi example, i get the following warning at the
>>> end,
>>> >>>> >> what must i do about this warning....thanks much for your help.
>>> >>>> >> Sitaraman
>>> >>>> >> inished in 20.82 seconds
>>> >>>> >> Java HotSpot(TM) 64-Bit Server VM warning: You have loaded
>>> library
>>> >>>> >> /usr/local/Software/hadoop-0.23.10/lib/native/libhadoop.so.1.0.0
>>> >>>> >> which
>>> >>>> >> might have disabled stack guard. The VM will try to fix the stack
>>> >>>> >> guard now.
>>> >>>> >> It's highly recommended that you fix the library with 'execstack
>>> -c
>>> >>>> >> <libfile>', or link it with '-z noexecstack'.
>>> >>>> >> 13/12/24 10:05:19 WARN util.NativeCodeLoader: Unable to load
>>> >>>> >> native-hadoop library for your platform... using builtin-java
>>> >>>> >> classes
>>> >>>> >> where applicable
>>> >>>> >> Estimated value of Pi is 3.14127500000000000000
>>> >>>> >> [sitaraman@localhost mapreduce]$
>>> >>>> >>
>>> >>>> >> On 12/23/13, Nitin Pawar <nitinpawar...@gmail.com> wrote:
>>> >>>> >> > Can you try starting the process as non root user.
>>> >>>> >> > Give proper permissions to the user and start it as a different
>>> >>>> >> > user.
>>> >>>> >> >
>>> >>>> >> > Thanks,
>>> >>>> >> > Nitin
>>> >>>> >> >
>>> >>>> >> >
>>> >>>> >> > On Mon, Dec 23, 2013 at 2:15 PM, Sitaraman Vilayannur <
>>> >>>> >> > vrsitaramanietfli...@gmail.com> wrote:
>>> >>>> >> >
>>> >>>> >> >> Hi,
>>> >>>> >> >>  When i attempt to start nodemanager i get the following
>>> error.
>>> >>>> >> >> Any
>>> >>>> >> >> help
>>> >>>> >> >> appreciated.   I was able to start resource manager datanode,
>>> >>>> namenode
>>> >>>> >> >> and
>>> >>>> >> >> secondarynamenode,
>>> >>>> >> >>
>>> >>>> >> >>
>>> >>>> >> >>    ./yarn-daemon.sh start nodemanager
>>> >>>> >> >> starting nodemanager, logging to
>>> >>>> >> >>
>>> >>>> >>
>>> >>>>
>>> /usr/local/Software/hadoop-0.23.10/logs/yarn-root-nodemanager-localhost.localdomain.out
>>> >>>> >> >> Unrecognized option: -jvm
>>> >>>> >> >> Error: Could not create the Java Virtual Machine.
>>> >>>> >> >> Error: A fatal exception has occurred. Program will exit.
>>> >>>> >> >> [root@localhost sbin]# emacs
>>> >>>> >> >>
>>> >>>> >>
>>> >>>>
>>> /usr/local/Software/hadoop-0.23.10/logs/yarn-root-nodemanager-localhost.localdomain.out
>>> >>>> >> >> &
>>> >>>> >> >> [4] 29004
>>> >>>> >> >> [root@localhost sbin]# jps
>>> >>>> >> >> 28402 SecondaryNameNode
>>> >>>> >> >> 30280 Jps
>>> >>>> >> >> 28299 DataNode
>>> >>>> >> >> 6729 Main
>>> >>>> >> >> 26044 ResourceManager
>>> >>>> >> >> 28197 NameNode
>>> >>>> >> >>
>>> >>>> >> >
>>> >>>> >> >
>>> >>>> >> >
>>> >>>> >> > --
>>> >>>> >> > Nitin Pawar
>>> >>>> >> >
>>> >>>> >>
>>> >>>> >
>>> >>>> >
>>> >>>> >
>>> >>>> > --
>>> >>>> > Nitin Pawar
>>> >>>> >
>>> >>>>
>>> >>>
>>> >>>
>>> >>>
>>> >>> --
>>> >>> Nitin Pawar
>>> >>>
>>> >>
>>> >
>>>
>>
>>
>>
>> --
>> Nitin Pawar
>>
>
>


-- 
Nitin Pawar

Reply via email to