Please set the compute nodes in slaves file at $HADOOP_HOME/etc/hadoop/slaves
run the following commands in $HADOOP_HOME/sbin to start the HDFS and Yarn Services hadoop-daemon.sh start namenode //start the namenode service hadoop-daemons.sh start datanode //start datanode on all nodes listed in slaves file yarn-daemon.sh start resourcemanager //start the resourcemanager yarn-daemons.sh start nodemanager // start nodemanager service on all nodes listed in slaves file On Tue, Nov 25, 2014 at 2:22 PM, Anand Murali <anand_vi...@yahoo.com> wrote: > Dear Alex: > > I am trying to install Hadoop-2.5.2 on Suse Enterprise Desktop 11 ONLY in > standalone/pseudo-distributed mode. Ambari needs a server. Now these are > the changes I have made in hadoop-env.sh based on Tom Whyte's text book > "Hadoop the definitive guide". > > export JAVA_HOME=/usr/lib64/jdk1.7.0_71/jdk7u71 > export HADOOP_HOME=/home/anand_vihar/hadoop > export PATH=:$PATH:$JAVA_HOME:$HADOOP_HOME/bin:$HADOOP_HOME/sbin > > All other variables are left un-touched as they are supposed to pick the > right defaults. Once having done this at > > $hadoop version > > Hadoop runs and shows version, which is first step successful the > > $hadoop namenode -format > > Is successful except for some warnings. I have set deafults in > core-site.xml, hdfs-site.xml and yarn-site.xml > > then > > $start-dfs.sh > > I get plenty of errors.. I am wondering if there is a clear cut install > procedure, or do you think Suse Desktop Enterprise 11 does not support > Hadoop. Reply welcome. > > Thanks > > Regards, > > Anand Murali. > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > Anand Murali > 11/7, 'Anand Vihar', Kandasamy St, Mylapore > Chennai - 600 004, India > Ph: (044)- 28474593/ 43526162 (voicemail) > > > On Tuesday, November 25, 2014 2:22 PM, AlexWang <wangxin...@gmail.com> > wrote: > > > Normally we only need to configure the environment variables in ~/.bashrc > or /etc/profile file, you can also configure the hadoop-env.sh file, they > are not in conflict. > I think hadoop-env.sh variables will override .bashrc variables. > For your question, you can try setting HDFS_CONF_DIR variables. Then try. > Cloudera hadoop installation you can use Cloudera Manager tool > > http://www.cloudera.com/content/cloudera/en/documentation/core/latest/topics/cm_ig_install_path_a.html > Install apache hadoop, unzip the tar.gz file and configure hadoop-related > configuration files and environment variables. > apache hadoop installation tools: http: //ambari.apache.org/ > > > On Nov 25, 2014, at 16:12, Anand Murali <anand_vi...@yahoo.com> wrote: > > Dear Alex: > > If I make changes to .bashrc, the above variables, will it not conflict > with hadoop-env.sh. And I was advised other then just JAVA_HOME, no other > environment variables should be set. Please advise. > > Thanks > > Anand Murali > 11/7, 'Anand Vihar', Kandasamy St, Mylapore > Chennai - 600 004, India > Ph: (044)- 28474593/ 43526162 (voicemail) > > > On Tuesday, November 25, 2014 1:23 PM, AlexWang <wangxin...@gmail.com> > wrote: > > > hadoop environment variable for example : > > echo " > export HADOOP_HOME=/usr/lib/hadoop > export HADOOP_HDFS_HOME=/usr/lib/hadoop-hdfs > export HADOOP_MAPRED_HOME=/usr/lib/hadoop-mapreduce > #export HADOOP_MAPRED_HOME=/usr/lib/hadoop-0.20-mapreduce > export HADOOP_COMMON_HOME=\${HADOOP_HOME} > export HADOOP_LIBEXEC_DIR=\${HADOOP_HOME}/libexec > export HADOOP_CONF_DIR=\${HADOOP_HOME}/etc/hadoop > *export HDFS_CONF_DIR=\${HADOOP_HOME}/etc/hadoop* > export HADOOP_YARN_HOME=/usr/lib/hadoop-yarn > export YARN_CONF_DIR=\${HADOOP_HOME}/etc/hadoop > export HADOOP_COMMON_LIB_NATIVE_DIR=\${HADOOP_HOME}/lib/native > export LD_LIBRARY_PATH=\${HADOOP_HOME}/lib/native > export HADOOP_OPTS=\"\${HADOOP_OPTS} > -Djava.library.path=\${HADOOP_HOME}/lib:\${LD_LIBRARY_PATH}\" > export PATH=\${HADOOP_HOME}/bin:\${HADOOP_HOME}/sbin:\$PATH > > ">> ~/.bashrc > > . ~/.bashrc > > > > > On Nov 24, 2014, at 21:25, Anand Murali <anand_vi...@yahoo.com> wrote: > > Dear All: > > After hadoop namenode -format I do the following with errors. > > anand_vihar@linux-v4vm:~/hadoop/etc/hadoop> hadoop start-dfs.sh > Error: Could not find or load main class start-dfs.sh > anand_vihar@linux-v4vm:~/hadoop/etc/hadoop> start-dfs.sh > Incorrect configuration: namenode address dfs.namenode.servicerpc-address > or dfs.namenode.rpc-address is not configured. > Starting namenodes on [2014-11-24 18:47:27,717 WARN [main] > util.NativeCodeLoader (NativeCodeLoader.java:<clinit>(62)) - Unable to load > native-hadoop library for your platform... using builtin-java classes where > applicable] > Error: Cannot find configuration directory: /etc/hadoop > Error: Cannot find configuration directory: /etc/hadoop > Starting secondary namenodes [2014-11-24 18:47:28,457 WARN [main] > util.NativeCodeLoader (NativeCodeLoader.java:<clinit>(62)) - Unable to load > native-hadoop library for your platform... using builtin-java classes where > applicable > 0.0.0.0] > Error: Cannot find configuration directory: /etc/hadoop > > But in my hadoop-env.sh I have set > > export JAVA_HOME=/usr/lib64/jdk1.7.1_71/jdk7u71 > export HADOOP_HOME=/anand_vihar/hadoop > export PATH=:PATH:$HADOOP_HOME/bin:$HADOOP_HOME/sbin:$HADOOP_HOME/share > > Would anyone know how to fix this problem. > > Thanks > > Regards, > > > Anand Murali > 11/7, 'Anand Vihar', Kandasamy St, Mylapore > Chennai - 600 004, India > Ph: (044)- 28474593/ 43526162 (voicemail) > > > On Monday, November 24, 2014 6:30 PM, Anand Murali <anand_vi...@yahoo.com> > wrote: > > > it works thanks > > Anand Murali > 11/7, 'Anand Vihar', Kandasamy St, Mylapore > Chennai - 600 004, India > Ph: (044)- 28474593/ 43526162 (voicemail) > > > On Monday, November 24, 2014 6:19 PM, Anand Murali <anand_vi...@yahoo.com> > wrote: > > > Ok. Many thanks I shall try. > > Anand Murali > 11/7, 'Anand Vihar', Kandasamy St, Mylapore > Chennai - 600 004, India > Ph: (044)- 28474593/ 43526162 (voicemail) > > > On Monday, November 24, 2014 6:13 PM, Rohith Sharma K S < > rohithsharm...@huawei.com> wrote: > > > The problem is with setting JAVA_HOME. There is .(Dot) before /usr which > cause append current directory. > export JAVA_HOME=*./**usr/lib64/jdk1.7.0_71/jdk7u71* > > *Do not use .(Dot) before /usr.* > > Thanks & Regards > Rohith Sharma K S > > This e-mail and its attachments contain confidential information from > HUAWEI, which is intended only for the person or entity whose address is > listed above. Any use of the information contained herein in any way > (including, but not limited to, total or partial disclosure, reproduction, > or dissemination) by persons other than the intended recipient(s) is > prohibited. If you receive this e-mail in error, please notify the sender > by phone or email immediately and delete it! > > *From:* Anand Murali [mailto:anand_vi...@yahoo.com <anand_vi...@yahoo.com> > ] > *Sent:* 24 November 2014 17:44 > *To:* user@hadoop.apache.org; user@hadoop.apache.org > *Subject:* Hadoop Installation Path problem > > Hi All: > > > I have done the follwoing in hadoop-env.sh > > export JAVA_HOME=./usr/lib64/jdk1.7.0_71/jdk7u71 > export HADOOP_HOME=/home/anand_vihar/hadoop > export PATH=:$PATH:$JAVA_HOME:$HADOOP_HOME/bin:$HADOOP_HOME/sbin > > Now when I run hadoop-env.sh and type hadoop version, I get this error. > > /home/anand_vihar/hadoop/bin/hadoop: line 133: > /home/anand_vihar/hadoop/etc/hadoop/usr/lib64/jdk1.7.0_71/jdk7u71/bin/java: > No such file or directory > /home/anand_vihar/hadoop/bin/hadoop: line 133: exec: > /home/anand_vihar/hadoop/etc/hadoop/usr/lib64/jdk1.7.0_71/jdk7u71/bin/java: > cannot execute: No such file or directory > > > Can somebody advise. I have asked this to many people, they all say the > obvious path problem, but where I cannot debug. This has become a show > stopper for me. Help most welcome. > > Thanks > > Regards > > > Anand Murali > 11/7, 'Anand Vihar', Kandasamy St, Mylapore > Chennai - 600 004, India > Ph: (044)- 28474593/ 43526162 (voicemail) > > > > > > > >