Hi the docs only have a property : mapreduce.input.fileinputformat.split.minsize (default value is 0) does it matter?
2013/3/14 Zheyi RONG <rongzh...@gmail.com> > Have you considered change mapred.max.split.size ? > As in: > http://stackoverflow.com/questions/9678180/change-file-split-size-in-hadoop > > Zheyi > > > On Thu, Mar 14, 2013 at 3:27 PM, YouPeng Yang > <yypvsxf19870...@gmail.com>wrote: > >> Hi >> >> >> I have done some tests in my Pseudo Mode(CDH4.1.2)with MV2 yarn,and >> : >> According to the doc: >> *mapreduce.jobtracker.address :*The host and port that the MapReduce >> job tracker runs at. If "local", then jobs are run in-process as a single >> map and reduce task. >> *mapreduce.job.maps (default value is 2)* :The default number of map >> tasks per job. Ignored when mapreduce.jobtracker.address is "local". >> >> I changed the mapreduce.jobtracker.address = Hadoop:50031. >> >> And then run the wordcount examples: >> hadoop jar hadoop-mapreduce-examples-2.0.0-cdh4.1.2.jar wordcount >> input output >> >> the output logs are as follows: >> .... >> Job Counters >> Launched map tasks=1 >> Launched reduce tasks=1 >> Data-local map tasks=1 >> Total time spent by all maps in occupied slots (ms)=60336 >> Total time spent by all reduces in occupied slots (ms)=63264 >> Map-Reduce Framework >> Map input records=5 >> Map output records=7 >> Map output bytes=56 >> Map output materialized bytes=76 >> .... >> >> i seem to does not work. >> >> I thought maybe my input file is small-just 5 records . is it right? >> >> regards >> >> >> >> >> >> >> >> 2013/3/14 Sai Sai <saigr...@yahoo.in> >> >>> >>> >>> In Pseudo Mode where is the setting to increase the number of mappers >>> or is this not possible. >>> Thanks >>> Sai >>> >> >> >