Thanks for your response Rasit. You may have missed a portion of my post.

> On a different note, when I attempt to pass params via -D I get a usage
message; when I use
> -jobconf the command goes through (and works in the case of
mapred.reduce.tasks=0 for
> example) but I get  a deprecation warning).

I'm using Hadoop 0.19.0 and -D is not working. Are you using version 0.19.0
as well?

John


On Wed, Feb 18, 2009 at 9:14 AM, Rasit OZDAS <rasitoz...@gmail.com> wrote:

> John, did you try -D option instead of -jobconf,
>
> I had -D option in my code, I changed it with -jobconf, this is what I get:
>
> ...
> ...
> Options:
>  -input    <path>     DFS input file(s) for the Map step
>  -output   <path>     DFS output directory for the Reduce step
>  -mapper   <cmd|JavaClassName>      The streaming command to run
>  -combiner <JavaClassName> Combiner has to be a Java class
>  -reducer  <cmd|JavaClassName>      The streaming command to run
>  -file     <file>     File/dir to be shipped in the Job jar file
>  -inputformat
> TextInputFormat(default)|SequenceFileAsTextInputFormat|JavaClassName
> Optional.
>  -outputformat TextOutputFormat(default)|JavaClassName  Optional.
>  -partitioner JavaClassName  Optional.
>  -numReduceTasks <num>  Optional.
>  -inputreader <spec>  Optional.
>  -cmdenv   <n>=<v>    Optional. Pass env.var to streaming commands
>  -mapdebug <path>  Optional. To run this script when a map task fails
>  -reducedebug <path>  Optional. To run this script when a reduce task fails
>
>  -verbose
>
> Generic options supported are
> -conf <configuration file>     specify an application configuration file
> -D <property=value>            use value for given property
> -fs <local|namenode:port>      specify a namenode
> -jt <local|jobtracker:port>    specify a job tracker
> -files <comma separated list of files>    specify comma separated files to
> be copied to the map reduce cluster
> -libjars <comma separated list of jars>    specify comma separated jar
> files
> to include in the classpath.
> -archives <comma separated list of archives>    specify comma separated
> archives to be unarchived on the compute machines.
>
> The general command line syntax is
> bin/hadoop command [genericOptions] [commandOptions]
>
> For more details about these options:
> Use $HADOOP_HOME/bin/hadoop jar build/hadoop-streaming.jar -info
>
>
>
> I think -jobconf is not used in v.0.19 .
>
> 2009/2/18 S D <sd.codewarr...@gmail.com>
>
> > I'm having trouble overriding the maximum number of map tasks that run on
> a
> > given machine in my cluster. The default value of
> > mapred.tasktracker.map.tasks.maximum is set to 2 in hadoop-default.xml.
> > When
> > running my job I passed
> >
> > -jobconf mapred.tasktracker.map.tasks.maximum=1
> >
> > to limit map tasks to one per machine but each machine was still
> allocated
> > 2
> > map tasks (simultaneously).  The only way I was able to guarantee a
> maximum
> > of one map task per machine was to change the value of the property in
> > hadoop-site.xml. This is unsatisfactory since I'll often be changing the
> > maximum on a per job basis. Any hints?
> >
> > On a different note, when I attempt to pass params via -D I get a usage
> > message; when I use -jobconf the command goes through (and works in the
> > case
> > of mapred.reduce.tasks=0 for example) but I get  a deprecation warning).
> >
> > Thanks,
> > John
> >
>
>
>
> --
> M. Raşit ÖZDAŞ
>

Reply via email to