in failed/killed task attempts, i see the following:

attempt_201012141048_0023_m_000000_0task_201012141048_0023_m_000000172.24.10.91FAILED
Too many fetch-failures
Last 4KB
Last 8KB
All
attempt_201012141048_0023_m_000000_1task_201012141048_0023_m_000000172.24.10.91FAILED
Too many fetch-failures
Last 4KB
Last 8KB
All
attempt_201012141048_0023_m_000001_0task_201012141048_0023_m_000001172.24.10.91FAILED
Too many fetch-failures
Last 4KB
Last 8KB
All
attempt_201012141048_0023_m_000001_1task_201012141048_0023_m_000001172.24.10.91FAILED
Too many fetch-failures
Last 4KB
Last 8KB
All
attempt_201012141048_0023_r_000000_0task_201012141048_0023_r_000000172.24.10.91FAILED
Shuffle Error: Exceeded MAX_FAILED_UNIQUE_FETCHES; bailing-out.
Last 4KB
Last 8KB
All
attempt_201012141048_0023_r_000000_1task_201012141048_0023_r_000000172.24.10.91FAILED
Shuffle Error: Exceeded MAX_FAILED_UNIQUE_FETCHES; bailing-out.
Last 4KB
Last 8KB
All
attempt_201012141048_0023_r_000000_2task_201012141048_0023_r_000000172.24.10.91FAILED
Shuffle Error: Exceeded MAX_FAILED_UNIQUE_FETCHES; bailing-out.
Last 4KB
Last 8KB
All
attempt_201012141048_0023_r_000000_3task_201012141048_0023_r_000000172.24.10.91FAILED
Shuffle Error: Exceeded MAX_FAILED_UNIQUE_FETCHES; bailing-out.



On Dec 20, 2010, at 11:01 PM, Adarsh Sharma wrote:

> Sean Curtis wrote:
>> just running a simple select count(1) from a table (using movielens as an 
>> example) doesnt seem to work for me.  anyone know why this doesnt work? im 
>> using hive trunk:
>> 
>> hive> select avg(rating) from movierating where movieid=43;
>> Total MapReduce jobs = 1
>> Launching Job 1 out of 1
>> Number of reduce tasks determined at compile time: 1
>> In order to change the average load for a reducer (in bytes):
>>  set hive.exec.reducers.bytes.per.reducer=<number>
>> In order to limit the maximum number of reducers:
>>  set hive.exec.reducers.max=<number>
>> In order to set a constant number of reducers:
>>  set mapred.reduce.tasks=<number>
>> Starting Job = job_201012141048_0023, Tracking URL = 
>> http://localhost:50030/jobdetails.jsp?jobid=job_201012141048_0023
>> Kill Command = /Users/Sean/dev/hadoop-0.20.2+737/bin/../bin/hadoop job  
>> -Dmapred.job.tracker=localhost:8021 -kill job_201012141048_0023
>> 2010-12-20 15:15:03,295 Stage-1 map = 0%,  reduce = 0%
>> 2010-12-20 15:15:09,420 Stage-1 map = 50%,  reduce = 0%
>> ... eventually fails after a couple of minutes with:
>> 
>> 2010-12-20 17:33:01,113 Stage-1 map = 100%,  reduce = 0%
>> 2010-12-20 17:33:32,182 Stage-1 map = 100%,  reduce = 100%
>> Ended Job = job_201012141048_0023 with errors
>> FAILED: Execution Error, return code 2 from 
>> org.apache.hadoop.hive.ql.exec.MapRedTask
>> hive> 
>> 
>> almost seems like the reduce task never starts. any help would be 
>> appreciated.
>> 
>> sean
> To know the root cause of the problem, got to Jobtracker web UI ( IP:50030) 
> and Check Job Tracker History at the bottom corresponding to this Job ID.
> 
> 
> Best Regards
> 
> Adarsh Sharma

Reply via email to