On 2/18/10 10:39 AM, Rob Stewart wrote:
Hi there,

I am using Hadoop 0.20.1 and I am trying to submit jobs to the cluster
as jars.

Here's what I'm trying to do:

 > hadoop jar $hadoopPath/hadoop-*-examples.jar join
-Dmapred.reduce.tasks=10 -inFormat
org.apache.hadoop.mapred.KeyValueTextInputFormat  -outKey
org.apache.hadoop.io.Text file1.dat file2.dat output.dat


However, my parameter setting of attempting to state 10 reducers for the
job is not being honoured by Hadoop, instead choosing some other number.
Where am I going wrong here? I do not want to have to change this value
in hadoop/conf.*.xml files, as I am attempting to show the expressive
power of Hadoop. Note: The power to specify the number or reducers in
possible in both Pig and Hive.


Thanks,


Rob Stewart

Rob:

It's possible that something inside the jar is calling JobConf.setNumReducers(x) after it parses the command line args. That would cause this type of behavior. I haven't looked at the source for the join example to confirm this, though.

Regards.
--
Eric Sammer
[email protected]
http://esammer.blogspot.com

Reply via email to