I see, John. I also use 0.19, just to note, -D option should come first, since it's one of generic options. I use it without any errors.
Cheers, Rasit 2009/2/18 S D <sd.codewarr...@gmail.com> > Thanks for your response Rasit. You may have missed a portion of my post. > > > On a different note, when I attempt to pass params via -D I get a usage > message; when I use > > -jobconf the command goes through (and works in the case of > mapred.reduce.tasks=0 for > > example) but I get a deprecation warning). > > I'm using Hadoop 0.19.0 and -D is not working. Are you using version 0.19.0 > as well? > > John > > > On Wed, Feb 18, 2009 at 9:14 AM, Rasit OZDAS <rasitoz...@gmail.com> wrote: > > > John, did you try -D option instead of -jobconf, > > > > I had -D option in my code, I changed it with -jobconf, this is what I > get: > > > > ... > > ... > > Options: > > -input <path> DFS input file(s) for the Map step > > -output <path> DFS output directory for the Reduce step > > -mapper <cmd|JavaClassName> The streaming command to run > > -combiner <JavaClassName> Combiner has to be a Java class > > -reducer <cmd|JavaClassName> The streaming command to run > > -file <file> File/dir to be shipped in the Job jar file > > -inputformat > > TextInputFormat(default)|SequenceFileAsTextInputFormat|JavaClassName > > Optional. > > -outputformat TextOutputFormat(default)|JavaClassName Optional. > > -partitioner JavaClassName Optional. > > -numReduceTasks <num> Optional. > > -inputreader <spec> Optional. > > -cmdenv <n>=<v> Optional. Pass env.var to streaming commands > > -mapdebug <path> Optional. To run this script when a map task fails > > -reducedebug <path> Optional. To run this script when a reduce task > fails > > > > -verbose > > > > Generic options supported are > > -conf <configuration file> specify an application configuration file > > -D <property=value> use value for given property > > -fs <local|namenode:port> specify a namenode > > -jt <local|jobtracker:port> specify a job tracker > > -files <comma separated list of files> specify comma separated files > to > > be copied to the map reduce cluster > > -libjars <comma separated list of jars> specify comma separated jar > > files > > to include in the classpath. > > -archives <comma separated list of archives> specify comma separated > > archives to be unarchived on the compute machines. > > > > The general command line syntax is > > bin/hadoop command [genericOptions] [commandOptions] > > > > For more details about these options: > > Use $HADOOP_HOME/bin/hadoop jar build/hadoop-streaming.jar -info > > > > > > > > I think -jobconf is not used in v.0.19 . > > > > 2009/2/18 S D <sd.codewarr...@gmail.com> > > > > > I'm having trouble overriding the maximum number of map tasks that run > on > > a > > > given machine in my cluster. The default value of > > > mapred.tasktracker.map.tasks.maximum is set to 2 in hadoop-default.xml. > > > When > > > running my job I passed > > > > > > -jobconf mapred.tasktracker.map.tasks.maximum=1 > > > > > > to limit map tasks to one per machine but each machine was still > > allocated > > > 2 > > > map tasks (simultaneously). The only way I was able to guarantee a > > maximum > > > of one map task per machine was to change the value of the property in > > > hadoop-site.xml. This is unsatisfactory since I'll often be changing > the > > > maximum on a per job basis. Any hints? > > > > > > On a different note, when I attempt to pass params via -D I get a usage > > > message; when I use -jobconf the command goes through (and works in the > > > case > > > of mapred.reduce.tasks=0 for example) but I get a deprecation > warning). > > > > > > Thanks, > > > John > > > > > > > > > > > -- > > M. Raşit ÖZDAŞ > > > -- M. Raşit ÖZDAŞ