For such small input, the only way you would see speed gains would be if your job was dominated by cpu time, and not i/o. Since word-count is mostly an i/o problem and your input size is quite small, you are seeing similar run times. 3 computers is better than 1 only if you need them.
On Apr 18, 2011, at 10:06 PM, praveenesh kumar wrote: > The input were 3 plain text files.. > > 1 file was around 665 KB and other 2 files were around 1.5 MB each.. > > Thanks, > Praveeenesh > > > > On Tue, Apr 19, 2011 at 10:27 AM, real great.. <greatness.hardn...@gmail.com >> wrote: > >> Whats your input size? >> >> On Tue, Apr 19, 2011 at 10:21 AM, praveenesh kumar <praveen...@gmail.com >>> wrote: >> >>> Hello everyone, >>> >>> I am new to hadoop... >>> I set up a hadoop cluster of 4 ubuntu systems. ( Hadoop 0.20.2) >>> and I am running the well known word count (gutenberg) example to test >> how >>> fast my hadoop is working.. >>> >>> But whenever I am running wordcount example..I am not able to see any >> much >>> processing time difference.. >>> On single node the wordcount is taking the same time.. and on cluster of >> 4 >>> systems also it is taking almost the same time.. >>> >>> Am I doing anything wrong here ?? >>> Can anyone explain me why its happening.. and how can I make maximum use >> of >>> my cluster ?? >>> >>> Thanks. >>> Praveenesh >>> >> >> >> >> -- >> Regards, >> R.V. >>