To add to Suresh's guideline since he may have missed providing a link, you can visit the CDH users community at https://groups.google.com/a/cloudera.org/forum/#!forum/cdh-user
On Fri, Mar 16, 2012 at 12:13 AM, Suresh Srinivas <sur...@hortonworks.com>wrote: > Guys, can you please take this up in CDH related mailing lists. > > On Thu, Mar 15, 2012 at 10:01 AM, Manu S <manupk...@gmail.com> wrote: > > > Because for large clusters we have to run namenode in a single node, > > datanode in another nodes > > So we can start namenode and jobtracker in master node and datanode n > > tasktracker in slave nodes > > > > For getting more clarity You can check the service status after starting > > > > Verify these: > > dfs.name.dir hdfs:hadoop drwx------ > > dfs.data.dir hdfs:hadoop drwx------ > > > > mapred.local.dir mapred:hadoop drwxr-xr-x > > > > Please follow each steps in this link > > https://ccp.cloudera.com/display/CDHDOC/CDH3+Deployment+on+a+Cluster > > On Mar 15, 2012 9:52 PM, "Manish Bhoge" <manishbh...@rocketmail.com> > > wrote: > > > > > Ys, I understand the order and I formatted namenode before starting > > > services. As I suspect there may be ownership and an access issue. Not > > able > > > to nail down issue exactly. I also have question why there are 2 routes > > to > > > start services. When we have start-all.sh script then why need to go to > > > init.d to start services?? > > > > > > > > > Thank you, > > > Manish > > > Sent from my BlackBerry, pls excuse typo > > > > > > -----Original Message----- > > > From: Manu S <manupk...@gmail.com> > > > Date: Thu, 15 Mar 2012 21:43:26 > > > To: <common-user@hadoop.apache.org>; <manishbh...@rocketmail.com> > > > Reply-To: common-user@hadoop.apache.org > > > Subject: Re: Issue when starting services on CDH3 > > > > > > Did you check the service status? > > > Is it like "dead, but pid exist"? > > > > > > Did you check the ownership and permissions for the > > > dfs.name.dir,dfs.data.dir,mapped.local.dir etc ? > > > > > > The order for starting daemons are like this: > > > 1 namenode > > > 2 datanode > > > 3 jobtracker > > > 4 tasktracker > > > > > > Did you format the namenode before starting? > > > On Mar 15, 2012 9:31 PM, "Manu S" <manupk...@gmail.com> wrote: > > > > > > > Dear manish > > > > Which daemons are not starting? > > > > > > > > On Mar 15, 2012 9:21 PM, "Manish Bhoge" <manishbh...@rocketmail.com> > > > > wrote: > > > > > > > > > > I have CDH3 installed in standalone mode. I have install all hadoop > > > > components. Now when I start services (namenode,secondary > namenode,job > > > > tracker,task tracker) I can start gracefully from /usr/lib/hadoop/ > > > > ./bin/start-all.sh. But when start the same servises from > > > > /etc/init.d/hadoop-0.20-* then I unable to start. Why? Now I want to > > > start > > > > Hue also which is in init.d that also I couldn't start. Here I > suspect > > > > authentication issue. Because all the services in init.d are under > root > > > > user and root group. Please suggest I am stuck here. I tried hive and > > it > > > > seems it running fine. > > > > > Thanks > > > > > Manish. > > > > > Sent from my BlackBerry, pls excuse typo > > > > > > > > > > > > > > > > > > -- Harsh J