Isn't this the same question you asked earlier today? I responded to the initial one that "-D mapred.task.timeout=18000000" shouldn't have a space after the D. It should be "-Dmapred.task.timeout=18000000"
And IIRC, these Hadoop parameters need to go before all of your other parameters. On Mon, Oct 22, 2012 at 4:54 PM, Joshi, Amit Krishna <[email protected]> wrote: > Hi, > > I am running PFPGrowth on several datasets and it works well for smaller ones > (< 5GB) > However, for the larger ones, I keep getting following timeout message. > > Task attempt_201210140938_0105_r_000000_0 failed to report status for 600 > seconds. Killing! > > Is there a way I can increase the timeout? > > I even tried passing these parameter but in vain: > -D mapred.task.timeout=18000000 -D mapred.child.java.opts=-Xmx4000m > > My input params are: -s 10000 -g 1000 -tc 8 -k 50 -method mapreduce > > Also, please suggest what would be the optimum value of g and k. > Number of features: > million > > > Thanks, > Amit
