Ok, so the problem is you have a machine with 16 CPUs and a
computation job that uses at most 3 of them. Is this right?

What is the Mahout task? Do you know that it has good multi-Hadoop
performance and tuning? What matters is that the Partitioner for the
Mahout code can separate the computations.



2010/11/30 Matthew Foley <ma...@yahoo-inc.com>:
> Here is a "recipe" for how to run multiple datanodes on a single server, 
> posted to this list on Sept. 15:
>        
> http://mail-archives.apache.org/mod_mbox/hadoop-common-user/201009.mbox/%3c8a898c33-dc4e-418c-adc0-5689d434b...@yahoo-inc.com%3e
>
> If you're having trouble getting multiple cores utilized,  this should help, 
> assuming the computation is well expressed in Map/Reduce.
> There is no point in running VMs in your situation, as it just adds overhead 
> without changing the amount of physical processor power
> available.
>
> --Matt
>
> On Nov 30, 2010, at 1:21 AM, beneo_7 wrote:
>
>> If you want to just use one machine, why do you want to use hadoop? Hadoop's
>> power lies in distributed computing. That being said, it is possible to use
>> hadoop on a single machine by using the pseudo-distributed mode (Read
>> http://hadoop.apache.org/common/docs/current/single_node_setup.html and
>> http://www.michael-noll.com/tutorials/running-hadoop-on-ubuntu-linux-single-node-cluster/).
>> If you are using just one machine, at least ensure that your machine has
>> lots of cores (8 core/16 cores would be great) to get benefit out of hadoop.
>>
>> I am not sure, but using virtual machines won't be helpful here as a virtual
>> machine is just an abstraction and not real hardware.
>
>
> thanks very much, i use the hadoop because the apache Mahout project need it 
> for clustering.
>
> my machine is only one and powerful, 16cores and 32g mem, but i have only 
> one, so i need configuration about multi nodes on one machine.
>
> i have used the "pseudo-distributed mode", however, the project always used 1 
> core,  the cpu freq always 100% ~ 103%, the time the execution is 4 hours, 
> it's too slow.
>
> i can not change the mahout project source code, the trunk always update 
> sometime, it's difficulty to solute confliction.
>
>
> is there any way to conf some slaves in one machine?
>
> At 2010-11-30 17:07:49,"Hari Sreekumar" <hsreeku...@clickable.com> wrote:
>
>> Hi beneo,
>>
>> If you want to just use one machine, why do you want to use hadoop? Hadoop's
>> power lies in distributed computing. That being said, it is possible to use
>> hadoop on a single machine by using the pseudo-distributed mode (Read
>> http://hadoop.apache.org/common/docs/current/single_node_setup.html and
>> http://www.michael-noll.com/tutorials/running-hadoop-on-ubuntu-linux-single-node-cluster/).
>> If you are using just one machine, at least ensure that your machine has
>> lots of cores (8 core/16 cores would be great) to get benefit out of hadoop.
>>
>> I am not sure, but using virtual machines won't be helpful here as a virtual
>> machine is just an abstraction and not real hardware.
>>
>> Cheers,
>> Hari
>>
>> 2010/11/30 beneo_7 <bene...@163.com>
>>
>>> i'm sorry, but, are you sure??
>>> At 2010-11-30 15:53:58,"rahul patodi" <patodira...@gmail.com> wrote:
>>>> you can create virtual machines on your single machine:
>>>> for you have to install sun virtual box(other tools are also available
>>> like
>>>> VMware)
>>>> now you can create as many virtual machine as you want
>>>> then you can create one master and all slaves
>>>>
>>>> -Thanks and Regards,
>>>> Rahul Patodi
>>>> Associate Software Engineer,
>>>> Impetus Infotech (India) Private Limited,
>>>> www.impetus.com
>>>> Mob:09907074413
>>>>
>>>> 2010/11/30 beneo_7 <bene...@163.com>
>>>>
>>>>> i have only one machine and it's powerful.
>>>>> so, i want the all the slaves and master on one machine?
>>>>>
>>>>> thx in advanced
>>>>>
>>>>
>>>>
>>>>
>>>> --
>>>
>>>
>
>



-- 
Lance Norskog
goks...@gmail.com

Reply via email to