Hi Tariq,

im not getting the right start..

On Thu, Aug 9, 2012 at 7:59 PM, Mohammad Tariq <donta...@gmail.com> wrote:

> Hello Rahul,
>
>    That's great. That's the best way to learn(I am doing the same :)
> ). Since the installation part is over, I would suggest to get
> yourself familiar with Hdfs and MapReduce first. Try to do basic
> filesystem operations using the Hdfs API and run the wordcount
> program, if you haven't done it yet. Then move ahead.
>
> Regards,
>     Mohammad Tariq
>
>
> On Thu, Aug 9, 2012 at 5:20 PM, rahul p <rahulpoolancha...@gmail.com>
> wrote:
> > Hi Tariq,
> >
> > I am also new to Hadoop trying to learn my self can anyone help me on the
> > same.
> > i have installed CDH3.
> >
> >
> >
> > On Thu, Aug 9, 2012 at 6:21 PM, Mohammad Tariq <donta...@gmail.com>
> wrote:
> >>
> >> Hello Anand,
> >>
> >>     Is there any specific reason behind not using ssh??
> >>
> >> Regards,
> >>     Mohammad Tariq
> >>
> >>
> >> On Thu, Aug 9, 2012 at 3:46 PM, anand sharma <anand2sha...@gmail.com>
> >> wrote:
> >> > Hi, i am just learning the Hadoop and i am setting the development
> >> > environment with CDH3 pseudo distributed mode without any ssh
> >> > cofiguration
> >> > in CentOS 6.2 . i can run the sample programs as usual but when i try
> >> > and
> >> > run namenode this is the error it logs...
> >> >
> >> > [hive@localhost ~]$ hadoop namenode
> >> > 12/08/09 20:56:57 INFO namenode.NameNode: STARTUP_MSG:
> >> > /************************************************************
> >> > STARTUP_MSG: Starting NameNode
> >> > STARTUP_MSG:   host = localhost.localdomain/127.0.0.1
> >> > STARTUP_MSG:   args = []
> >> > STARTUP_MSG:   version = 0.20.2-cdh3u4
> >> > STARTUP_MSG:   build =
> >> > file:///data/1/tmp/topdir/BUILD/hadoop-0.20.2-cdh3u4
> >> > -r 214dd731e3bdb687cb55988d3f47dd9e248c5690; compiled by 'root' on Mon
> >> > May
> >> > 7 14:01:59 PDT 2012
> >> > ************************************************************/
> >> > 12/08/09 20:56:57 INFO jvm.JvmMetrics: Initializing JVM Metrics with
> >> > processName=NameNode, sessionId=null
> >> > 12/08/09 20:56:57 INFO metrics.NameNodeMetrics: Initializing
> >> > NameNodeMeterics using context
> >> > object:org.apache.hadoop.metrics.spi.NoEmitMetricsContext
> >> > 12/08/09 20:56:57 INFO util.GSet: VM type       = 64-bit
> >> > 12/08/09 20:56:57 INFO util.GSet: 2% max memory = 17.77875 MB
> >> > 12/08/09 20:56:57 INFO util.GSet: capacity      = 2^21 = 2097152
> entries
> >> > 12/08/09 20:56:57 INFO util.GSet: recommended=2097152, actual=2097152
> >> > 12/08/09 20:56:57 INFO namenode.FSNamesystem: fsOwner=hive
> (auth:SIMPLE)
> >> > 12/08/09 20:56:57 INFO namenode.FSNamesystem: supergroup=supergroup
> >> > 12/08/09 20:56:57 INFO namenode.FSNamesystem:
> isPermissionEnabled=false
> >> > 12/08/09 20:56:57 INFO namenode.FSNamesystem:
> >> > dfs.block.invalidate.limit=1000
> >> > 12/08/09 20:56:57 INFO namenode.FSNamesystem:
> isAccessTokenEnabled=false
> >> > accessKeyUpdateInterval=0 min(s), accessTokenLifetime=0 min(s)
> >> > 12/08/09 20:56:57 INFO metrics.FSNamesystemMetrics: Initializing
> >> > FSNamesystemMetrics using context
> >> > object:org.apache.hadoop.metrics.spi.NoEmitMetricsContext
> >> > 12/08/09 20:56:57 ERROR namenode.FSNamesystem: FSNamesystem
> >> > initialization
> >> > failed.
> >> > java.io.FileNotFoundException:
> >> > /var/lib/hadoop-0.20/cache/hadoop/dfs/name/in_use.lock (Permission
> >> > denied)
> >> > at java.io.RandomAccessFile.open(Native Method)
> >> > at java.io.RandomAccessFile.<init>(RandomAccessFile.java:216)
> >> > at
> >> >
> >> >
> org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.tryLock(Storage.java:614)
> >> > at
> >> >
> >> >
> org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.lock(Storage.java:591)
> >> > at
> >> >
> >> >
> org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.analyzeStorage(Storage.java:449)
> >> > at
> >> >
> >> >
> org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:304)
> >> > at
> >> >
> >> >
> org.apache.hadoop.hdfs.server.namenode.FSDirectory.loadFSImage(FSDirectory.java:110)
> >> > at
> >> >
> >> >
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.initialize(FSNamesystem.java:372)
> >> > at
> >> >
> >> >
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.<init>(FSNamesystem.java:335)
> >> > at
> >> >
> >> >
> org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:271)
> >> > at
> >> >
> org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:467)
> >> > at
> >> >
> >> >
> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1330)
> >> > at
> >> >
> org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1339)
> >> > 12/08/09 20:56:57 ERROR namenode.NameNode:
> >> > java.io.FileNotFoundException:
> >> > /var/lib/hadoop-0.20/cache/hadoop/dfs/name/in_use.lock (Permission
> >> > denied)
> >> > at java.io.RandomAccessFile.open(Native Method)
> >> > at java.io.RandomAccessFile.<init>(RandomAccessFile.java:216)
> >> > at
> >> >
> >> >
> org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.tryLock(Storage.java:614)
> >> > at
> >> >
> >> >
> org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.lock(Storage.java:591)
> >> > at
> >> >
> >> >
> org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.analyzeStorage(Storage.java:449)
> >> > at
> >> >
> >> >
> org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:304)
> >> > at
> >> >
> >> >
> org.apache.hadoop.hdfs.server.namenode.FSDirectory.loadFSImage(FSDirectory.java:110)
> >> > at
> >> >
> >> >
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.initialize(FSNamesystem.java:372)
> >> > at
> >> >
> >> >
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.<init>(FSNamesystem.java:335)
> >> > at
> >> >
> >> >
> org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:271)
> >> > at
> >> >
> org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:467)
> >> > at
> >> >
> >> >
> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1330)
> >> > at
> >> >
> org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1339)
> >> >
> >> > 12/08/09 20:56:57 INFO namenode.NameNode: SHUTDOWN_MSG:
> >> > /************************************************************
> >> > SHUTDOWN_MSG: Shutting down NameNode at localhost.localdomain/
> 127.0.0.1
> >> > ************************************************************/
> >> >
> >> >
> >
> >
>

Reply via email to