Dear JM,

It's correct. 

The Hbase version is 0.94.2 and the hadoop version is 0.20.2 / 1.04。

We test this on both hadoop version 0.20.2 and 1.04. 

The error still there. 

Thanks a lot



Best Regards / 商祺
郭伟 Guo Wei
----------------------------------------------------- 
南京西桥科技有限公司 
Western Bridge Tech Ltd.,  Nanjing

南京市玄武区花园路8号一号楼511
No. 511, Building 1, No. 8, Hua Yuan Road

Xuanwu District, Nanjing, PR China

Email: [email protected]

Tel: +86 25 8528 4900 (Operator) 
Mobile: +86 138 1589 8257
Fax: +86 25 8528 4980

Weibo: http://weibo.com/guowee
Web: http://www.wbkit.com
----------------------------------------------------- 
WesternBridge Tech: Professional software service provider. Professional is 
MANNER as well CAPABILITY.

On 2013-3-26, at 下午9:18, Jean-Marc Spaggiari <[email protected]> wrote:

> Hi,
> 
> So basically, you have one job which is reading  from A and writing to
> B, and one wich is reading from A and writing to C, and  the 2 jobs
> are running at the same time. Is that correct? Are you able to
> reproduce that each time you are running the job? Which HBased and
> Hadoop versions are you running?
> 
> JM
> 
> 2013/3/26 GuoWei <[email protected]>:
>> Dear,
>> 
>> When I run two MR Jobs which will read same Hbase table and write to another 
>> same Hbase table at the same time. The result is one job successful 
>> finished. And another job crashed. And The following shows the error log.
>> 
>> Please help me to find out why ?
>> 
>> 
>> <2013-03-25 15:50:34,026> <INFO > <org.apache.hadoop.mapred.JobClient> -  
>> map 0% reduce 0%(JobClient.java:monitorAndPrintJob:1301)
>> <2013-03-25 15:50:36,096> <WARN > <org.apache.hadoop.mapred.Task> - Could 
>> not find output size (Task.java:calculateOutputSize:948)
>> org.apache.hadoop.util.DiskChecker$DiskErrorException: Could not find 
>> output/file.out in any of the configured local directories
>>        at 
>> org.apache.hadoop.fs.LocalDirAllocator$AllocatorPerContext.getLocalPathToRead(LocalDirAllocator.java:429)
>>        at 
>> org.apache.hadoop.fs.LocalDirAllocator.getLocalPathToRead(LocalDirAllocator.java:160)
>>        at 
>> org.apache.hadoop.mapred.MapOutputFile.getOutputFile(MapOutputFile.java:56)
>>        at org.apache.hadoop.mapred.Task.calculateOutputSize(Task.java:944)
>>        at org.apache.hadoop.mapred.Task.sendLastUpdate(Task.java:924)
>>        at org.apache.hadoop.mapred.Task.done(Task.java:875)
>>        at org.apache.hadoop.mapred.MapTask.run(MapTask.java:374)
>>        at 
>> org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:212)
>> <2013-03-25 15:50:36,100> <INFO > <org.apache.hadoop.mapred.LocalJobRunner> 
>> - (LocalJobRunner.java:statusUpdate:321)
>> <2013-03-25 15:50:36,102> <INFO > <org.apache.hadoop.mapred.Task> - Task 
>> 'attempt_local_0001_m_000000_0' done.(Task.java:sendDone:959)
>> <2013-03-25 15:50:36,111> <WARN > 
>> <org.apache.hadoop.mapred.FileOutputCommitter> - Output path is null in 
>> cleanup(FileOutputCommitter.java:cleanupJob:100)
>> <2013-03-25 15:50:36,111> <WARN > <org.apache.hadoop.mapred.LocalJobRunner> 
>> - job_local_0001(LocalJobRunner.java:run:298)
>> org.apache.hadoop.util.DiskChecker$DiskErrorException: Could not find 
>> output/file.out in any of the configured local directories
>>        at 
>> org.apache.hadoop.fs.LocalDirAllocator$AllocatorPerContext.getLocalPathToRead(LocalDirAllocator.java:429)
>>        at 
>> org.apache.hadoop.fs.LocalDirAllocator.getLocalPathToRead(LocalDirAllocator.java:160)
>>        at 
>> org.apache.hadoop.mapred.MapOutputFile.getOutputFile(MapOutputFile.java:56)
>>        at 
>> org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:236)
>> <2013-03-25 15:50:37,029> <INFO > <org.apache.hadoop.mapred.JobClient> -  
>> map 100% reduce 0%(JobClient.java:monitorAndPrintJob:1301)
>> <2013-03-25 15:50:37,030> <INFO > <org.apache.hadoop.mapred.JobClient> - Job 
>> complete: job_local_0001(JobClient.java:monitorAndPrintJob:1356)
>> <2013-03-25 15:50:37,031> <INFO > <org.apache.hadoop.mapred.JobClient> - 
>> Counters: 15(Counters.java:log:585)
>> <2013-03-25 15:50:37,031> <INFO > <org.apache.hadoop.mapred.JobClient> -   
>> File Input Format Counters (Counters.java:log:587)
>> <2013-03-25 15:50:37,031> <INFO > <org.apache.hadoop.mapred.JobClient> -     
>> Bytes Read=0(Counters.java:log:589)
>> <2013-03-25 15:50:37,032> <INFO > <org.apache.hadoop.mapred.JobClient> -   
>> FileSystemCounters(Counters.java:log:587)
>> <2013-03-25 15:50:37,032> <INFO > <org.apache.hadoop.mapred.JobClient> -     
>> FILE_BYTES_READ=10294950(Counters.java:log:589)
>> <2013-03-25 15:50:37,033> <INFO > <org.apache.hadoop.mapred.JobClient> -     
>> FILE_BYTES_WRITTEN=10432139(Counters.java:log:589)
>> <2013-03-25 15:50:37,033> <INFO > <org.apache.hadoop.mapred.JobClient> -   
>> Map-Reduce Framework(Counters.java:log:587)
>> <2013-03-25 15:50:37,033> <INFO > <org.apache.hadoop.mapred.JobClient> -     
>> Map output materialized bytes=4006(Counters.java:log:589)
>> <2013-03-25 15:50:37,034> <INFO > <org.apache.hadoop.mapred.JobClient> -     
>> Combine output records=0(Counters.java:log:589)
>> <2013-03-25 15:50:37,034> <INFO > <org.apache.hadoop.mapred.JobClient> -     
>> Map input records=500(Counters.java:log:589)
>> <2013-03-25 15:50:37,035> <INFO > <org.apache.hadoop.mapred.JobClient> -     
>> Physical memory (bytes) snapshot=0(Counters.java:log:589)
>> <2013-03-25 15:50:37,035> <INFO > <org.apache.hadoop.mapred.JobClient> -     
>> Spilled Records=500(Counters.java:log:589)
>> <2013-03-25 15:50:37,035> <INFO > <org.apache.hadoop.mapred.JobClient> -     
>> Map output bytes=3000(Counters.java:log:589)
>> <2013-03-25 15:50:37,036> <INFO > <org.apache.hadoop.mapred.JobClient> -     
>> Total committed heap usage (bytes)=202702848(Counters.java:log:589)
>> <2013-03-25 15:50:37,036> <INFO > <org.apache.hadoop.mapred.JobClient> -     
>> CPU time spent (ms)=0(Counters.java:log:589)
>> <2013-03-25 15:50:37,037> <INFO > <org.apache.hadoop.mapred.JobClient> -     
>> Virtual memory (bytes) snapshot=0(Counters.java:log:589)
>> <2013-03-25 15:50:37,037> <INFO > <org.apache.hadoop.mapred.JobClient> -     
>> SPLIT_RAW_BYTES=105(Counters.java:log:589)
>> <2013-03-25 15:50:37,038> <INFO > <org.apache.hadoop.mapred.JobClient> -     
>> Map output records=500(Counters.java:log:589)
>> <2013-03-25 15:50:37,038> <INFO > <org.apache.hadoop.mapred.JobClient> -     
>> Combine input
>> 
>> Thanks a lot.
>> 
>> 
>> 
>> Best Regards
>> 
>> Weibo: http://weibo.com/guowee
>> Web: http://www.wbkit.com
>> -----------------------------------------------------
>> WesternBridge Tech: Professional software service provider. Professional is 
>> MANNER as well CAPABILITY.
>> 
> 

Reply via email to