I would change "mapred.tasktracker.reduce.tasks.maximum" to one. With your
setting

On May 2, 2011 8:48 AM, "baran cakici" <barancak...@gmail.com> wrote:
> without job;
>
> CPU Usage = 0%
> Memory = 585 MB (2GB Ram)
>
> Baran
> 2011/5/2 baran cakici <barancak...@gmail.com>
>
>> CPU Usage = 95-100%
>> Memory = 650-850 MB (2GB Ram)
>>
>> Baran
>>
>>
>> 2011/5/2 James Seigel <ja...@tynt.com>
>>
>>> If you have windows and cygwin you probably don't have a lot if memory
>>> left at 2 gig.
>>>
>>> Pull up system monitor on the data nodes and check for free memory
>>> when you have you jobs running. I bet it is quite low.
>>>
>>> I am not a windows guy so I can't take you much farther.
>>>
>>> James
>>>
>>> Sent from my mobile. Please excuse the typos.
>>>
>>> On 2011-05-02, at 8:32 AM, baran cakici <barancak...@gmail.com> wrote:
>>>
>>> > yes, I am running under cygwin on my datanodes too. OS of Datanodes
are
>>> > Windows as well.
>>> >
>>> > What can I do exactly for a better Performance. I changed
>>> > mapred.child.java.opts to default value.How can I solve this
"swapping"
>>> > problem?
>>> >
>>> > PS: I dont have a chance to get Slaves(Celeron 2GHz) with Liniux OS.
>>> >
>>> > thanks, both of you
>>> >
>>> > Regards,
>>> >
>>> > Baran
>>> > 2011/5/2 Richard Nadeau <strout...@gmail.com>
>>> >
>>> >> Are you running under cygwin on your data nodes as well? That is
>>> certain to
>>> >> cause performance problems. As James suggested, swapping to disk is
>>> going
>>> >> to
>>> >> be a killer, running on Windows with Celeron processors only
compounds
>>> the
>>> >> problem. The Celeron processor is also sub-optimal for CPU intensive
>>> tasks
>>> >>
>>> >> Rick
>>> >>
>>> >> On Apr 28, 2011 9:22 AM, "baran cakici" <barancak...@gmail.com>
wrote:
>>> >>> Hi Everyone,
>>> >>>
>>> >>> I have a Cluster with one Master(JobTracker and NameNode - Intel
>>> Core2Duo
>>> >> 2
>>> >>> GB Ram) and four Slaves(Datanode and Tasktracker - Celeron 2 GB
Ram).
>>> My
>>> >>> Inputdata are between 2GB-10GB and I read Inputdata in MapReduce
line
>>> by
>>> >>> line. Now, I try to accelerate my System(Benchmark), but I'm not
sure,
>>> if
>>> >> my
>>> >>> Configuration is correctly. Can you please just look, if it is ok?
>>> >>>
>>> >>> -mapred-site.xml
>>> >>>
>>> >>> <property>
>>> >>> <name>mapred.job.tracker</name>
>>> >>> <value>apple:9001</value>
>>> >>> </property>
>>> >>>
>>> >>> <property>
>>> >>> <name>mapred.child.java.opts</name>
>>> >>> <value>-Xmx512m -server</value>
>>> >>> </property>
>>> >>>
>>> >>> <property>
>>> >>> <name>mapred.job.tracker.handler.count</name>
>>> >>> <value>2</value>
>>> >>> </property>
>>> >>>
>>> >>> <property>
>>> >>> <name>mapred.local.dir</name>
>>> >>>
>>> >>
>>>
<value>/cygwin/usr/local/hadoop-datastore/hadoop-Baran/mapred/local</value>
>>> >>> </property>
>>> >>>
>>> >>> <property>
>>> >>> <name>mapred.map.tasks</name>
>>> >>> <value>1</value>
>>> >>> </property>
>>> >>>
>>> >>> <property>
>>> >>> <name>mapred.reduce.tasks</name>
>>> >>> <value>4</value>
>>> >>> </property>
>>> >>>
>>> >>> <property>
>>> >>> <name>mapred.submit.replication</name>
>>> >>> <value>2</value>
>>> >>> </property>
>>> >>>
>>> >>> <property>
>>> >>> <name>mapred.system.dir</name>
>>> >>>
>>> >>
>>> >>
>>>
<value>/cygwin/usr/local/hadoop-datastore/hadoop-Baran/mapred/system</value>
>>> >>> </property>
>>> >>>
>>> >>> <property>
>>> >>> <name>mapred.tasktracker.indexcache.mb</name>
>>> >>> <value>10</value>
>>> >>> </property>
>>> >>>
>>> >>> <property>
>>> >>> <name>mapred.tasktracker.map.tasks.maximum</name>
>>> >>> <value>1</value>
>>> >>> </property>
>>> >>>
>>> >>> <property>
>>> >>> <name>mapred.tasktracker.reduce.tasks.maximum</name>
>>> >>> <value>4</value>
>>> >>> </property>
>>> >>>
>>> >>> <property>
>>> >>> <name>mapred.temp.dir</name>
>>> >>>
>>> >>
>>>
<value>/cygwin/usr/local/hadoop-datastore/hadoop-Baran/mapred/temp</value>
>>> >>> </property>
>>> >>>
>>> >>> <property>
>>> >>> <name>webinterface.private.actions</name>
>>> >>> <value>true</value>
>>> >>> </property>
>>> >>>
>>> >>> <property>
>>> >>> <name>mapred.reduce.slowstart.completed.maps</name>
>>> >>> <value>0.01</value>
>>> >>> </property>
>>> >>>
>>> >>> -hdfs-site.xml
>>> >>>
>>> >>> <property>
>>> >>> <name>dfs.block.size</name>
>>> >>> <value>268435456</value>
>>> >>> </property>
>>> >>> PS: I extended dfs.block.size, because I won 50% better performance
>>> with
>>> >>> this change.
>>> >>>
>>> >>> I am waiting for your comments...
>>> >>>
>>> >>> Regards,
>>> >>>
>>> >>> Baran
>>> >>
>>>
>>
>>

Reply via email to