The firewall was run at system startup, I think there was a /etc/sysconfig/iptables file present which triggered the firewall. I don't currently have access to any centos 5 machines so I can't easily check.
On Thu, Apr 16, 2009 at 6:54 PM, jason hadoop <jason.had...@gmail.com>wrote: > The kickstart script was something that the operations staff was using to > initialize new machines, I never actually saw the script, just figured out > that there was a firewall in place. > > > > On Thu, Apr 16, 2009 at 1:28 PM, Mithila Nagendra <mnage...@asu.edu>wrote: > >> Jason: the kickstart script - was it something you wrote or is it run when >> the system turns on? >> Mithila >> >> On Thu, Apr 16, 2009 at 1:06 AM, Mithila Nagendra <mnage...@asu.edu> >> wrote: >> >> > Thanks Jason! Will check that out. >> > Mithila >> > >> > >> > On Thu, Apr 16, 2009 at 5:23 AM, jason hadoop <jason.had...@gmail.com >> >wrote: >> > >> >> Double check that there is no firewall in place. >> >> At one point a bunch of new machines were kickstarted and placed in a >> >> cluster and they all failed with something similar. >> >> It turned out the kickstart script turned enabled the firewall with a >> rule >> >> that blocked ports in the 50k range. >> >> It took us a while to even think to check that was not a part of our >> >> normal >> >> machine configuration >> >> >> >> On Wed, Apr 15, 2009 at 11:04 AM, Mithila Nagendra <mnage...@asu.edu> >> >> wrote: >> >> >> >> > Hi Aaron >> >> > I will look into that thanks! >> >> > >> >> > I spoke to the admin who overlooks the cluster. He said that the >> gateway >> >> > comes in to the picture only when one of the nodes communicates with >> a >> >> node >> >> > outside of the cluster. But in my case the communication is carried >> out >> >> > between the nodes which all belong to the same cluster. >> >> > >> >> > Mithila >> >> > >> >> > On Wed, Apr 15, 2009 at 8:59 PM, Aaron Kimball <aa...@cloudera.com> >> >> wrote: >> >> > >> >> > > Hi, >> >> > > >> >> > > I wrote a blog post a while back about connecting nodes via a >> gateway. >> >> > See >> >> > > >> >> > >> >> >> http://www.cloudera.com/blog/2008/12/03/securing-a-hadoop-cluster-through-a-gateway/ >> >> > > >> >> > > This assumes that the client is outside the gateway and all >> >> > > datanodes/namenode are inside, but the same principles apply. >> You'll >> >> just >> >> > > need to set up ssh tunnels from every datanode to the namenode. >> >> > > >> >> > > - Aaron >> >> > > >> >> > > >> >> > > On Wed, Apr 15, 2009 at 10:19 AM, Ravi Phulari < >> >> rphul...@yahoo-inc.com >> >> > >wrote: >> >> > > >> >> > >> Looks like your NameNode is down . >> >> > >> Verify if hadoop process are running ( jps should show you all >> java >> >> > >> running process). >> >> > >> If your hadoop process are running try restarting your hadoop >> process >> >> . >> >> > >> I guess this problem is due to your fsimage not being correct . >> >> > >> You might have to format your namenode. >> >> > >> Hope this helps. >> >> > >> >> >> > >> Thanks, >> >> > >> -- >> >> > >> Ravi >> >> > >> >> >> > >> >> >> > >> On 4/15/09 10:15 AM, "Mithila Nagendra" <mnage...@asu.edu> wrote: >> >> > >> >> >> > >> The log file runs into thousands of line with the same message >> being >> >> > >> displayed every time. >> >> > >> >> >> > >> On Wed, Apr 15, 2009 at 8:10 PM, Mithila Nagendra < >> mnage...@asu.edu> >> >> > >> wrote: >> >> > >> >> >> > >> > The log file : hadoop-mithila-datanode-node19.log.2009-04-14 has >> >> the >> >> > >> > following in it: >> >> > >> > >> >> > >> > 2009-04-14 10:08:11,499 INFO org.apache.hadoop.dfs.DataNode: >> >> > >> STARTUP_MSG: >> >> > >> > /************************************************************ >> >> > >> > STARTUP_MSG: Starting DataNode >> >> > >> > STARTUP_MSG: host = node19/127.0.0.1 >> >> > >> > STARTUP_MSG: args = [] >> >> > >> > STARTUP_MSG: version = 0.18.3 >> >> > >> > STARTUP_MSG: build = >> >> > >> > >> https://svn.apache.org/repos/asf/hadoop/core/branches/branch-0.18-r >> >> > >> > 736250; compiled by 'ndaley' on Thu Jan 22 23:12:08 UTC 2009 >> >> > >> > ************************************************************/ >> >> > >> > 2009-04-14 10:08:12,915 INFO org.apache.hadoop.ipc.Client: >> Retrying >> >> > >> connect >> >> > >> > to server: node18/192.168.0.18:54310. Already tried 0 time(s). >> >> > >> > 2009-04-14 10:08:13,925 INFO org.apache.hadoop.ipc.Client: >> Retrying >> >> > >> connect >> >> > >> > to server: node18/192.168.0.18:54310. Already tried 1 time(s). >> >> > >> > 2009-04-14 10:08:14,935 INFO org.apache.hadoop.ipc.Client: >> Retrying >> >> > >> connect >> >> > >> > to server: node18/192.168.0.18:54310. Already tried 2 time(s). >> >> > >> > 2009-04-14 10:08:15,945 INFO org.apache.hadoop.ipc.Client: >> Retrying >> >> > >> connect >> >> > >> > to server: node18/192.168.0.18:54310. Already tried 3 time(s). >> >> > >> > 2009-04-14 10:08:16,955 INFO org.apache.hadoop.ipc.Client: >> Retrying >> >> > >> connect >> >> > >> > to server: node18/192.168.0.18:54310. Already tried 4 time(s). >> >> > >> > 2009-04-14 10:08:17,965 INFO org.apache.hadoop.ipc.Client: >> Retrying >> >> > >> connect >> >> > >> > to server: node18/192.168.0.18:54310. Already tried 5 time(s). >> >> > >> > 2009-04-14 10:08:18,975 INFO org.apache.hadoop.ipc.Client: >> Retrying >> >> > >> connect >> >> > >> > to server: node18/192.168.0.18:54310. Already tried 6 time(s). >> >> > >> > 2009-04-14 10:08:19,985 INFO org.apache.hadoop.ipc.Client: >> Retrying >> >> > >> connect >> >> > >> > to server: node18/192.168.0.18:54310. Already tried 7 time(s). >> >> > >> > 2009-04-14 10:08:20,995 INFO org.apache.hadoop.ipc.Client: >> Retrying >> >> > >> connect >> >> > >> > to server: node18/192.168.0.18:54310. Already tried 8 time(s). >> >> > >> > 2009-04-14 10:08:22,005 INFO org.apache.hadoop.ipc.Client: >> Retrying >> >> > >> connect >> >> > >> > to server: node18/192.168.0.18:54310. Already tried 9 time(s). >> >> > >> > 2009-04-14 10:08:22,008 INFO org.apache.hadoop.ipc.RPC: Server >> at >> >> > >> node18/ >> >> > >> > 192.168.0.18:54310 not available yet, Zzzzz... >> >> > >> > 2009-04-14 10:08:24,025 INFO org.apache.hadoop.ipc.Client: >> Retrying >> >> > >> connect >> >> > >> > to server: node18/192.168.0.18:54310. Already tried 0 time(s). >> >> > >> > 2009-04-14 10:08:25,035 INFO org.apache.hadoop.ipc.Client: >> Retrying >> >> > >> connect >> >> > >> > to server: node18/192.168.0.18:54310. Already tried 1 time(s). >> >> > >> > 2009-04-14 10:08:26,045 INFO org.apache.hadoop.ipc.Client: >> Retrying >> >> > >> connect >> >> > >> > to server: node18/192.168.0.18:54310. Already tried 2 time(s). >> >> > >> > 2009-04-14 10:08:27,055 INFO org.apache.hadoop.ipc.Client: >> Retrying >> >> > >> connect >> >> > >> > to server: node18/192.168.0.18:54310. Already tried 3 time(s). >> >> > >> > 2009-04-14 10:08:28,065 INFO org.apache.hadoop.ipc.Client: >> Retrying >> >> > >> connect >> >> > >> > to server: node18/192.168.0.18:54310. Already tried 4 time(s). >> >> > >> > 2009-04-14 10:08:29,075 INFO org.apache.hadoop.ipc.Client: >> Retrying >> >> > >> connect >> >> > >> > to server: node18/192.168.0.18:54310. Already tried 5 time(s). >> >> > >> > 2009-04-14 10:08:30,085 INFO org.apache.hadoop.ipc.Client: >> Retrying >> >> > >> connect >> >> > >> > to server: node18/192.168.0.18:54310. Already tried 6 time(s). >> >> > >> > 2009-04-14 10:08:31,095 INFO org.apache.hadoop.ipc.Client: >> Retrying >> >> > >> connect >> >> > >> > to server: node18/192.168.0.18:54310. Already tried 7 time(s). >> >> > >> > 2009-04-14 10:08:32,105 INFO org.apache.hadoop.ipc.Client: >> Retrying >> >> > >> connect >> >> > >> > to server: node18/192.168.0.18:54310. Already tried 8 time(s). >> >> > >> > 2009-04-14 10:08:33,115 INFO org.apache.hadoop.ipc.Client: >> Retrying >> >> > >> connect >> >> > >> > to server: node18/192.168.0.18:54310. Already tried 9 time(s). >> >> > >> > 2009-04-14 10:08:33,116 INFO org.apache.hadoop.ipc.RPC: Server >> at >> >> > >> node18/ >> >> > >> > 192.168.0.18:54310 not available yet, Zzzzz... >> >> > >> > 2009-04-14 10:08:35,135 INFO org.apache.hadoop.ipc.Client: >> Retrying >> >> > >> connect >> >> > >> > to server: node18/192.168.0.18:54310. Already tried 0 time(s). >> >> > >> > 2009-04-14 10:08:36,145 INFO org.apache.hadoop.ipc.Client: >> Retrying >> >> > >> connect >> >> > >> > to server: node18/192.168.0.18:54310. Already tried 1 time(s). >> >> > >> > 2009-04-14 10:08:37,155 INFO org.apache.hadoop.ipc.Client: >> Retrying >> >> > >> connect >> >> > >> > to server: node18/192.168.0.18:54310. Already tried 2 time(s). >> >> > >> > >> >> > >> > >> >> > >> > Hmmm I still cant figure it out.. >> >> > >> > >> >> > >> > Mithila >> >> > >> > >> >> > >> > >> >> > >> > On Tue, Apr 14, 2009 at 10:22 PM, Mithila Nagendra < >> >> mnage...@asu.edu >> >> > >> >wrote: >> >> > >> > >> >> > >> >> Also, Would the way the port is accessed change if all these >> node >> >> are >> >> > >> >> connected through a gateway? I mean in the hadoop-site.xml >> file? >> >> The >> >> > >> Ubuntu >> >> > >> >> systems we worked with earlier didnt have a gateway. >> >> > >> >> Mithila >> >> > >> >> >> >> > >> >> On Tue, Apr 14, 2009 at 9:48 PM, Mithila Nagendra < >> >> mnage...@asu.edu >> >> > >> >wrote: >> >> > >> >> >> >> > >> >>> Aaron: Which log file do I look into - there are alot of them. >> >> Here >> >> > s >> >> > >> >>> what the error looks like: >> >> > >> >>> [mith...@node19:~]$ cd hadoop >> >> > >> >>> [mith...@node19:~/hadoop]$ bin/hadoop dfs -ls >> >> > >> >>> 09/04/14 10:09:29 INFO ipc.Client: Retrying connect to server: >> >> > node18/ >> >> > >> >>> 192.168.0.18:54310. Already tried 0 time(s). >> >> > >> >>> 09/04/14 10:09:30 INFO ipc.Client: Retrying connect to server: >> >> > node18/ >> >> > >> >>> 192.168.0.18:54310. Already tried 1 time(s). >> >> > >> >>> 09/04/14 10:09:31 INFO ipc.Client: Retrying connect to server: >> >> > node18/ >> >> > >> >>> 192.168.0.18:54310. Already tried 2 time(s). >> >> > >> >>> 09/04/14 10:09:32 INFO ipc.Client: Retrying connect to server: >> >> > node18/ >> >> > >> >>> 192.168.0.18:54310. Already tried 3 time(s). >> >> > >> >>> 09/04/14 10:09:33 INFO ipc.Client: Retrying connect to server: >> >> > node18/ >> >> > >> >>> 192.168.0.18:54310. Already tried 4 time(s). >> >> > >> >>> 09/04/14 10:09:34 INFO ipc.Client: Retrying connect to server: >> >> > node18/ >> >> > >> >>> 192.168.0.18:54310. Already tried 5 time(s). >> >> > >> >>> 09/04/14 10:09:35 INFO ipc.Client: Retrying connect to server: >> >> > node18/ >> >> > >> >>> 192.168.0.18:54310. Already tried 6 time(s). >> >> > >> >>> 09/04/14 10:09:36 INFO ipc.Client: Retrying connect to server: >> >> > node18/ >> >> > >> >>> 192.168.0.18:54310. Already tried 7 time(s). >> >> > >> >>> 09/04/14 10:09:37 INFO ipc.Client: Retrying connect to server: >> >> > node18/ >> >> > >> >>> 192.168.0.18:54310. Already tried 8 time(s). >> >> > >> >>> 09/04/14 10:09:38 INFO ipc.Client: Retrying connect to server: >> >> > node18/ >> >> > >> >>> 192.168.0.18:54310. Already tried 9 time(s). >> >> > >> >>> Bad connection to FS. command aborted. >> >> > >> >>> >> >> > >> >>> Node19 is a slave and Node18 is the master. >> >> > >> >>> >> >> > >> >>> Mithila >> >> > >> >>> >> >> > >> >>> >> >> > >> >>> >> >> > >> >>> On Tue, Apr 14, 2009 at 8:53 PM, Aaron Kimball < >> >> aa...@cloudera.com >> >> > >> >wrote: >> >> > >> >>> >> >> > >> >>>> Are there any error messages in the log files on those nodes? >> >> > >> >>>> - Aaron >> >> > >> >>>> >> >> > >> >>>> On Tue, Apr 14, 2009 at 9:03 AM, Mithila Nagendra < >> >> > mnage...@asu.edu> >> >> > >> >>>> wrote: >> >> > >> >>>> >> >> > >> >>>> > I ve drawn a blank here! Can't figure out what s wrong with >> >> the >> >> > >> ports. >> >> > >> >>>> I >> >> > >> >>>> > can >> >> > >> >>>> > ssh between the nodes but cant access the DFS from the >> slaves >> >> - >> >> > >> says >> >> > >> >>>> "Bad >> >> > >> >>>> > connection to DFS". Master seems to be fine. >> >> > >> >>>> > Mithila >> >> > >> >>>> > >> >> > >> >>>> > On Tue, Apr 14, 2009 at 4:28 AM, Mithila Nagendra < >> >> > >> mnage...@asu.edu> >> >> > >> >>>> > wrote: >> >> > >> >>>> > >> >> > >> >>>> > > Yes I can.. >> >> > >> >>>> > > >> >> > >> >>>> > > >> >> > >> >>>> > > On Mon, Apr 13, 2009 at 5:12 PM, Jim Twensky < >> >> > >> jim.twen...@gmail.com >> >> > >> >>>> > >wrote: >> >> > >> >>>> > > >> >> > >> >>>> > >> Can you ssh between the nodes? >> >> > >> >>>> > >> >> >> > >> >>>> > >> -jim >> >> > >> >>>> > >> >> >> > >> >>>> > >> On Mon, Apr 13, 2009 at 6:49 PM, Mithila Nagendra < >> >> > >> >>>> mnage...@asu.edu> >> >> > >> >>>> > >> wrote: >> >> > >> >>>> > >> >> >> > >> >>>> > >> > Thanks Aaron. >> >> > >> >>>> > >> > Jim: The three clusters I setup had ubuntu running on >> >> them >> >> > and >> >> > >> >>>> the dfs >> >> > >> >>>> > >> was >> >> > >> >>>> > >> > accessed at port 54310. The new cluster which I ve >> setup >> >> has >> >> > >> Red >> >> > >> >>>> Hat >> >> > >> >>>> > >> Linux >> >> > >> >>>> > >> > release 7.2 (Enigma)running on it. Now when I try to >> >> access >> >> > >> the >> >> > >> >>>> dfs >> >> > >> >>>> > from >> >> > >> >>>> > >> > one >> >> > >> >>>> > >> > of the slaves i get the following response: dfs cannot >> be >> >> > >> >>>> accessed. >> >> > >> >>>> > When >> >> > >> >>>> > >> I >> >> > >> >>>> > >> > access the DFS throught the master there s no problem. >> So >> >> I >> >> > >> feel >> >> > >> >>>> there >> >> > >> >>>> > a >> >> > >> >>>> > >> > problem with the port. Any ideas? I did check the list >> of >> >> > >> slaves, >> >> > >> >>>> it >> >> > >> >>>> > >> looks >> >> > >> >>>> > >> > fine to me. >> >> > >> >>>> > >> > >> >> > >> >>>> > >> > Mithila >> >> > >> >>>> > >> > >> >> > >> >>>> > >> > >> >> > >> >>>> > >> > >> >> > >> >>>> > >> > >> >> > >> >>>> > >> > On Mon, Apr 13, 2009 at 2:58 PM, Jim Twensky < >> >> > >> >>>> jim.twen...@gmail.com> >> >> > >> >>>> > >> > wrote: >> >> > >> >>>> > >> > >> >> > >> >>>> > >> > > Mithila, >> >> > >> >>>> > >> > > >> >> > >> >>>> > >> > > You said all the slaves were being utilized in the 3 >> >> node >> >> > >> >>>> cluster. >> >> > >> >>>> > >> Which >> >> > >> >>>> > >> > > application did you run to test that and what was >> your >> >> > input >> >> > >> >>>> size? >> >> > >> >>>> > If >> >> > >> >>>> > >> you >> >> > >> >>>> > >> > > tried the word count application on a 516 MB input >> file >> >> on >> >> > >> both >> >> > >> >>>> > >> cluster >> >> > >> >>>> > >> > > setups, than some of your nodes in the 15 node >> cluster >> >> may >> >> > >> not >> >> > >> >>>> be >> >> > >> >>>> > >> running >> >> > >> >>>> > >> > > at >> >> > >> >>>> > >> > > all. Generally, one map job is assigned to each >> input >> >> > split >> >> > >> and >> >> > >> >>>> if >> >> > >> >>>> > you >> >> > >> >>>> > >> > are >> >> > >> >>>> > >> > > running your cluster with the defaults, the splits >> are >> >> 64 >> >> > MB >> >> > >> >>>> each. I >> >> > >> >>>> > >> got >> >> > >> >>>> > >> > > confused when you said the Namenode seemed to do all >> >> the >> >> > >> work. >> >> > >> >>>> Can >> >> > >> >>>> > you >> >> > >> >>>> > >> > > check >> >> > >> >>>> > >> > > conf/slaves and make sure you put the names of all >> task >> >> > >> >>>> trackers >> >> > >> >>>> > >> there? I >> >> > >> >>>> > >> > > also suggest comparing both clusters with a larger >> >> input >> >> > >> size, >> >> > >> >>>> say >> >> > >> >>>> > at >> >> > >> >>>> > >> > least >> >> > >> >>>> > >> > > 5 GB, to really see a difference. >> >> > >> >>>> > >> > > >> >> > >> >>>> > >> > > Jim >> >> > >> >>>> > >> > > >> >> > >> >>>> > >> > > On Mon, Apr 13, 2009 at 4:17 PM, Aaron Kimball < >> >> > >> >>>> aa...@cloudera.com> >> >> > >> >>>> > >> > wrote: >> >> > >> >>>> > >> > > >> >> > >> >>>> > >> > > > in hadoop-*-examples.jar, use "randomwriter" to >> >> generate >> >> > >> the >> >> > >> >>>> data >> >> > >> >>>> > >> and >> >> > >> >>>> > >> > > > "sort" >> >> > >> >>>> > >> > > > to sort it. >> >> > >> >>>> > >> > > > - Aaron >> >> > >> >>>> > >> > > > >> >> > >> >>>> > >> > > > On Sun, Apr 12, 2009 at 9:33 PM, Pankil Doshi < >> >> > >> >>>> > forpan...@gmail.com> >> >> > >> >>>> > >> > > wrote: >> >> > >> >>>> > >> > > > >> >> > >> >>>> > >> > > > > Your data is too small I guess for 15 clusters >> ..So >> >> it >> >> > >> >>>> might be >> >> > >> >>>> > >> > > overhead >> >> > >> >>>> > >> > > > > time of these clusters making your total MR jobs >> >> more >> >> > >> time >> >> > >> >>>> > >> consuming. >> >> > >> >>>> > >> > > > > I guess you will have to try with larger set of >> >> data.. >> >> > >> >>>> > >> > > > > >> >> > >> >>>> > >> > > > > Pankil >> >> > >> >>>> > >> > > > > On Sun, Apr 12, 2009 at 6:54 PM, Mithila >> Nagendra < >> >> > >> >>>> > >> mnage...@asu.edu> >> >> > >> >>>> > >> > > > > wrote: >> >> > >> >>>> > >> > > > > >> >> > >> >>>> > >> > > > > > Aaron >> >> > >> >>>> > >> > > > > > >> >> > >> >>>> > >> > > > > > That could be the issue, my data is just 516MB >> - >> >> > >> wouldn't >> >> > >> >>>> this >> >> > >> >>>> > >> see >> >> > >> >>>> > >> > a >> >> > >> >>>> > >> > > > bit >> >> > >> >>>> > >> > > > > of >> >> > >> >>>> > >> > > > > > speed up? >> >> > >> >>>> > >> > > > > > Could you guide me to the example? I ll run my >> >> > cluster >> >> > >> on >> >> > >> >>>> it >> >> > >> >>>> > and >> >> > >> >>>> > >> > see >> >> > >> >>>> > >> > > > what >> >> > >> >>>> > >> > > > > I >> >> > >> >>>> > >> > > > > > get. Also for my program I had a java timer >> >> running >> >> > to >> >> > >> >>>> record >> >> > >> >>>> > >> the >> >> > >> >>>> > >> > > time >> >> > >> >>>> > >> > > > > > taken >> >> > >> >>>> > >> > > > > > to complete execution. Does Hadoop have an >> >> inbuilt >> >> > >> timer? >> >> > >> >>>> > >> > > > > > >> >> > >> >>>> > >> > > > > > Mithila >> >> > >> >>>> > >> > > > > > >> >> > >> >>>> > >> > > > > > On Mon, Apr 13, 2009 at 1:13 AM, Aaron Kimball >> < >> >> > >> >>>> > >> aa...@cloudera.com >> >> > >> >>>> > >> > > >> >> > >> >>>> > >> > > > > wrote: >> >> > >> >>>> > >> > > > > > >> >> > >> >>>> > >> > > > > > > Virtually none of the examples that ship >> with >> >> > Hadoop >> >> > >> >>>> are >> >> > >> >>>> > >> designed >> >> > >> >>>> > >> > > to >> >> > >> >>>> > >> > > > > > > showcase its speed. Hadoop's speedup comes >> from >> >> > its >> >> > >> >>>> ability >> >> > >> >>>> > to >> >> > >> >>>> > >> > > > process >> >> > >> >>>> > >> > > > > > very >> >> > >> >>>> > >> > > > > > > large volumes of data (starting around, say, >> >> tens >> >> > of >> >> > >> GB >> >> > >> >>>> per >> >> > >> >>>> > >> job, >> >> > >> >>>> > >> > > and >> >> > >> >>>> > >> > > > > > going >> >> > >> >>>> > >> > > > > > > up in orders of magnitude from there). So if >> >> you >> >> > are >> >> > >> >>>> timing >> >> > >> >>>> > >> the >> >> > >> >>>> > >> > pi >> >> > >> >>>> > >> > > > > > > calculator (or something like that), its >> >> results >> >> > >> won't >> >> > >> >>>> > >> > necessarily >> >> > >> >>>> > >> > > be >> >> > >> >>>> > >> > > > > > very >> >> > >> >>>> > >> > > > > > > consistent. If a job doesn't have enough >> >> fragments >> >> > >> of >> >> > >> >>>> data >> >> > >> >>>> > to >> >> > >> >>>> > >> > > > allocate >> >> > >> >>>> > >> > > > > > one >> >> > >> >>>> > >> > > > > > > per each node, some of the nodes will also >> just >> >> go >> >> > >> >>>> unused. >> >> > >> >>>> > >> > > > > > > >> >> > >> >>>> > >> > > > > > > The best example for you to run is to use >> >> > >> randomwriter >> >> > >> >>>> to >> >> > >> >>>> > fill >> >> > >> >>>> > >> up >> >> > >> >>>> > >> > > > your >> >> > >> >>>> > >> > > > > > > cluster with several GB of random data and >> then >> >> > run >> >> > >> the >> >> > >> >>>> sort >> >> > >> >>>> > >> > > program. >> >> > >> >>>> > >> > > > > If >> >> > >> >>>> > >> > > > > > > that doesn't scale up performance from 3 >> nodes >> >> to >> >> > >> 15, >> >> > >> >>>> then >> >> > >> >>>> > >> you've >> >> > >> >>>> > >> > > > > > > definitely >> >> > >> >>>> > >> > > > > > > got something strange going on. >> >> > >> >>>> > >> > > > > > > >> >> > >> >>>> > >> > > > > > > - Aaron >> >> > >> >>>> > >> > > > > > > >> >> > >> >>>> > >> > > > > > > >> >> > >> >>>> > >> > > > > > > On Sun, Apr 12, 2009 at 8:39 AM, Mithila >> >> Nagendra >> >> > < >> >> > >> >>>> > >> > > mnage...@asu.edu> >> >> > >> >>>> > >> > > > > > > wrote: >> >> > >> >>>> > >> > > > > > > >> >> > >> >>>> > >> > > > > > > > Hey all >> >> > >> >>>> > >> > > > > > > > I recently setup a three node hadoop >> cluster >> >> and >> >> > >> ran >> >> > >> >>>> an >> >> > >> >>>> > >> > examples >> >> > >> >>>> > >> > > on >> >> > >> >>>> > >> > > > > it. >> >> > >> >>>> > >> > > > > > > It >> >> > >> >>>> > >> > > > > > > > was pretty fast, and all the three nodes >> were >> >> > >> being >> >> > >> >>>> used >> >> > >> >>>> > (I >> >> > >> >>>> > >> > > checked >> >> > >> >>>> > >> > > > > the >> >> > >> >>>> > >> > > > > > > log >> >> > >> >>>> > >> > > > > > > > files to make sure that the slaves are >> >> > utilized). >> >> > >> >>>> > >> > > > > > > > >> >> > >> >>>> > >> > > > > > > > Now I ve setup another cluster consisting >> of >> >> 15 >> >> > >> >>>> nodes. I >> >> > >> >>>> > ran >> >> > >> >>>> > >> > the >> >> > >> >>>> > >> > > > same >> >> > >> >>>> > >> > > > > > > > example, but instead of speeding up, the >> >> > >> map-reduce >> >> > >> >>>> task >> >> > >> >>>> > >> seems >> >> > >> >>>> > >> > to >> >> > >> >>>> > >> > > > > take >> >> > >> >>>> > >> > > > > > > > forever! The slaves are not being used for >> >> some >> >> > >> >>>> reason. >> >> > >> >>>> > This >> >> > >> >>>> > >> > > second >> >> > >> >>>> > >> > > > > > > cluster >> >> > >> >>>> > >> > > > > > > > has a lower, per node processing power, >> but >> >> > should >> >> > >> >>>> that >> >> > >> >>>> > make >> >> > >> >>>> > >> > any >> >> > >> >>>> > >> > > > > > > > difference? >> >> > >> >>>> > >> > > > > > > > How can I ensure that the data is being >> >> mapped >> >> > to >> >> > >> all >> >> > >> >>>> the >> >> > >> >>>> > >> > nodes? >> >> > >> >>>> > >> > > > > > > Presently, >> >> > >> >>>> > >> > > > > > > > the only node that seems to be doing all >> the >> >> > work >> >> > >> is >> >> > >> >>>> the >> >> > >> >>>> > >> Master >> >> > >> >>>> > >> > > > node. >> >> > >> >>>> > >> > > > > > > > >> >> > >> >>>> > >> > > > > > > > Does 15 nodes in a cluster increase the >> >> network >> >> > >> cost? >> >> > >> >>>> What >> >> > >> >>>> > >> can >> >> > >> >>>> > >> > I >> >> > >> >>>> > >> > > do >> >> > >> >>>> > >> > > > > to >> >> > >> >>>> > >> > > > > > > > setup >> >> > >> >>>> > >> > > > > > > > the cluster to function more efficiently? >> >> > >> >>>> > >> > > > > > > > >> >> > >> >>>> > >> > > > > > > > Thanks! >> >> > >> >>>> > >> > > > > > > > Mithila Nagendra >> >> > >> >>>> > >> > > > > > > > Arizona State University >> >> > >> >>>> > >> > > > > > > > >> >> > >> >>>> > >> > > > > > > >> >> > >> >>>> > >> > > > > > >> >> > >> >>>> > >> > > > > >> >> > >> >>>> > >> > > > >> >> > >> >>>> > >> > > >> >> > >> >>>> > >> > >> >> > >> >>>> > >> >> >> > >> >>>> > > >> >> > >> >>>> > > >> >> > >> >>>> > >> >> > >> >>>> >> >> > >> >>> >> >> > >> >>> >> >> > >> >> >> >> > >> > >> >> > >> >> >> > >> >> >> > >> Ravi >> >> > >> -- >> >> > >> >> >> > >> >> >> > > >> >> > >> >> >> >> >> >> >> >> -- >> >> Alpha Chapters of my book on Hadoop are available >> >> http://www.apress.com/book/view/9781430219422 >> >> >> > >> > >> > > > > -- > Alpha Chapters of my book on Hadoop are available > http://www.apress.com/book/view/9781430219422 > -- Alpha Chapters of my book on Hadoop are available http://www.apress.com/book/view/9781430219422