For such small input, the only way you would see speed gains would be if your 
job was dominated
by cpu time, and not i/o. Since word-count is mostly an i/o problem and your 
input size is quite small, you are seeing similar run times. 3 computers is 
better than 1
only if you need them. 

On Apr 18, 2011, at 10:06 PM, praveenesh kumar wrote:

> The input were  3  plain text files..
> 
> 1 file was around 665 KB and other 2 files were around 1.5 MB each..
> 
> Thanks,
> Praveeenesh
> 
> 
> 
> On Tue, Apr 19, 2011 at 10:27 AM, real great.. <greatness.hardn...@gmail.com
>> wrote:
> 
>> Whats your input size?
>> 
>> On Tue, Apr 19, 2011 at 10:21 AM, praveenesh kumar <praveen...@gmail.com
>>> wrote:
>> 
>>> Hello everyone,
>>> 
>>> I am new to hadoop...
>>> I set up a  hadoop cluster of 4 ubuntu systems. ( Hadoop 0.20.2)
>>> and I am running the well known word count (gutenberg) example to test
>> how
>>> fast my hadoop is working..
>>> 
>>> But whenever I am running wordcount example..I am not able to see any
>> much
>>> processing time difference..
>>> On single node the wordcount is taking the same time.. and on cluster of
>> 4
>>> systems also it is taking almost the same time..
>>> 
>>> Am I  doing anything wrong here ??
>>> Can anyone explain me why its happening.. and how can I make maximum use
>> of
>>> my cluster ??
>>> 
>>> Thanks.
>>> Praveenesh
>>> 
>> 
>> 
>> 
>> --
>> Regards,
>> R.V.
>> 

Reply via email to