Hello Mates: Thanks to everyone for their help so far. I have learnt a lot and have now done single and pseudo mode. I have a hadoop cluster but I ran jps on the master node and slave node but not all process are started
master: 22160 NameNode 22716 Jps 22458 JobTracker slave: 32195 Jps I also checked the logs and I see files for all the datanodes, jobtracker, namenode, secondarynamenode, and tasktracker. The tasktracker has one slave node log missing. The namenode formatted correctly. I set the values for below so I'm not sure if I need more. My cluster is 11 nodes (1 master, 10 slaves). I do not have permission to access root only my directory so hadoop is installed in there. I can ssh to the slaves properly. * fs.default.name, dfs.name.dir, dfs.data.dir, mapred.job.tracker, mapred.system.dir It also gave errors regarding: * it cannot find the hadoop-daemon.sh file but I can see it /home/my-user/hadoop-0.20.2_cluster/bin/hadoop-daemon.sh: line 40: cd: /home/my-user/hadoop-0.20.2_cluster/bin: No such file or directory * it has the wrong path for the hadoop-config.sh so which parameter sets this field?? /home/my-user/hadoop-0.20.2_cluster/bin/hadoop-daemon.sh: line 42: /home/my-user/hadoop-0.20.2_cluster/hadoop-config.sh: No such file or directory * not being able to create the log directory on the same slave node that doesn't have its tasktracker, which parameters should be used to set the log directory? The same slave node which is giving problems also has: Usage: hadoop-daemon.sh [--config <conf-dir>] [--hosts hostlistfile] (start|stop) <hadoop-command> <args...> Thanks for your help. Cheers, Tamara