hi Ricky, for installing CDH3 you can refer this tutorial: http://cloudera-tutorial.blogspot.com/2010/11/running-cloudera-in-distributed-mode.html all the steps in this tutorial are well tested.(*in case of any query please leave a comment*)
On Wed, Nov 24, 2010 at 11:48 AM, rahul patodi <patodira...@gmail.com>wrote: > hi Hary, > when i try to start hadoop daemons by /usr/lib/hadoop# bin/start-dfs.sh on > name node it is giving this error:"*May not run daemons as root. Please > specify HADOOP_NAMENODE_USER"(*same for other daemons*)* > but when i try to start it using "*/etc/init.d/hadoop-0.20-namenode start" > *it* *gets start successfully* ** > * > *whats the reason behind that? > * > > On Wed, Nov 24, 2010 at 10:04 AM, Hari Sreekumar <hsreeku...@clickable.com > > wrote: > >> Hi Ricky, >> >> Yes, that's how it is meant to be. The machine where you run >> start-dfs.sh will become the namenode, and the machine whihc you specify >> in >> you "masters" file becomes the secondary namenode. >> >> Hari >> >> On Wed, Nov 24, 2010 at 2:13 AM, Ricky Ho <rickyphyl...@yahoo.com> wrote: >> >> > Thanks for pointing me to the right command. I am using the CDH3 >> > distribution. >> > I figure out no matter what I put in the masters file, it always start >> the >> > NamedNode at the machine where I issue the "start-all.sh" command. And >> > always >> > start a SecondaryNamedNode in all other machines. Any clue ? >> > >> > >> > Rgds, >> > Ricky >> > >> > -----Original Message----- >> > From: Hari Sreekumar [mailto:hsreeku...@clickable.com] >> > Sent: Tuesday, November 23, 2010 10:25 AM >> > To: common-user@hadoop.apache.org >> > Subject: Re: Is there a single command to start the whole cluster in >> CDH3 ? >> > >> > Hi Ricky, >> > >> > Which hadoop version are you using? I am using hadoop-0.20.2 >> apache >> > version, and I generally just run the $HADOOP_HOME/bin/start-dfs.sh and >> > start-mapred.sh script on my master node. If passwordless ssh is >> > configured, >> > this script will start the required services on each node. You shouldn't >> > have to start the services on each node individually. The secondary >> > namenode >> > is specified in the conf/masters file. The node where you call the >> > start-*.sh script becomes the namenode(for start-dfs) or jobtracker(for >> > start-mapred). The node mentioned in the masters file becomes the 2ndary >> > namenode, and the datanodes and tasktrackers are the nodes which are >> > mentioned in the slaves file. >> > >> > HTH, >> > Hari >> > >> > On Tue, Nov 23, 2010 at 11:43 PM, Ricky Ho <rickyphyl...@yahoo.com> >> wrote: >> > >> > > I setup the cluster configuration in "masters", "slaves", >> > "core-site.xml", >> > > "hdfs-site.xml", "mapred-site.xml" and copy to all the machines. >> > > >> > > And I login to one of the machines and use the following to start the >> > > cluster. >> > > for service in /etc/init.d/hadoop-0.20-*; do sudo $service start; done >> > > >> > > I expect this command will SSH to all the other machines (based on the >> > > "master" >> > > and "slaves" files) to start the corresponding daemons, but obviously >> it >> > is >> > > not >> > > doing that in my setup. >> > > >> > > Am I missing something in my setup ? >> > > >> > > Also, where do I specify where the Secondary Name Node is run. >> > > >> > > Rgds, >> > > Ricky >> > > >> > > >> > > >> > > >> > > >> > >> > >> > >> > >> > > > > -- > -Thanks and Regards, > Rahul Patodi > Associate Software Engineer, > Impetus Infotech (India) Private Limited, > www.impetus.com > Mob:09907074413 > > -- -Thanks and Regards, Rahul Patodi Associate Software Engineer, Impetus Infotech (India) Private Limited, www.impetus.com Mob:09907074413