Hi,

I am trying to iterate over 4.5m rows with about 1000 regions and an average row size of at least .5MB. Here is the end screen after the map run:

Hadoop job_200801101958_0001 on lv1-xen-pdc-2 User: root Job Name: building index Job File: /tmp/hadoop-root/mapred/system/job_200801101958_0001/job.xml Status: Running Started at: Thu Jan 10 20:05:11 PST 2008 Running for: 59mins, 8sec Black-listed TaskTrackers: 1 ________________________________________________________________________________________________________________________________________________________ Kind % Complete Num Tasks Pending Running Complete Killed Failed/Killed Task Attempts map 100.00% 995 0 0 995 0 7 / 36 reduce 1.57% 1 0 1 0 0 1 / 0 Counter Map Reduce Total Job Counters Launched map tasks 0 0 1,038 Launched reduce tasks 0 0 2 Map-Reduce Framework Map input records 4,454,969 0 4,454,969 Map output records 4,454,969 0 4,454,969 Map input bytes 0 0 0 Map output bytes 246,332,525,224 0 246,332,525,224 Combine input records 0 0 0 Combine output records 0 0 0 Reduce input groups 0 0 0 Reduce input records 0 0 0 Reduce output records 0 0 0 ________________________________________________________________________________________________________________________________________________________ Map Completion Graph - close [EMBED] ________________________________________________________________________________________________________________________________________________________ Reduce Completion Graph - close [EMBED]
    __________________________________________________________

Then it sort of ends. Here are the errors I got on the name node:

# /usr/local/hadoop/bin/hadoop jar /webapps/worldlingo/maindex.jar com.worldlingo.hadoop.mapred.index.BuildSearchIndex -m 1024 -r 1 -indexDir /pdcindex

org.apache.hadoop.hbase.UnknownScannerException: org.apache.hadoop.hbase.UnknownScannerException: 2472756350024972361 at org.apache.hadoop.hbase.HRegionServer.close(HRegionServer.java:1494)
       at sun.reflect.GeneratedMethodAccessor9.invoke(Unknown Source)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
       at java.lang.reflect.Method.invoke(Method.java:585)
       at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:401)
       at org.apache.hadoop.ipc.Server$Handler.run(Server.java:892)

at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:39) at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27)
       at java.lang.reflect.Constructor.newInstance(Constructor.java:494)
at org.apache.hadoop.hbase.RemoteExceptionHandler.decodeRemoteException(RemoteExceptionHandler.java:82) at org.apache.hadoop.hbase.HTable$ClientScanner.close(HTable.java:1099) at com.worldlingo.hadoop.mapred.index.TableInputFormat$TableRecordReader.close(TableInputFormat.java:65) at org.apache.hadoop.mapred.MapTask$TrackedRecordReader.close(MapTask.java:155)
       at org.apache.hadoop.mapred.MapTask.run(MapTask.java:212)
at org.apache.hadoop.mapred.TaskTracker$Child.main(TaskTracker.java:2043) org.apache.hadoop.util.DiskChecker$DiskErrorException: Could not find any valid local directory for task_200801101958_0001_m_000345_0/spill0.out at org.apache.hadoop.fs.LocalDirAllocator$AllocatorPerContext.getLocalPathForWrite(LocalDirAllocator.java:313) at org.apache.hadoop.fs.LocalDirAllocator.getLocalPathForWrite(LocalDirAllocator.java:124) at org.apache.hadoop.mapred.MapOutputFile.getSpillFileForWrite(MapOutputFile.java:88) at org.apache.hadoop.mapred.MapTask$MapOutputBuffer.sortAndSpillToDisk(MapTask.java:405) at org.apache.hadoop.mapred.MapTask$MapOutputBuffer.collect(MapTask.java:387)
       at com.worldlingo.hadoop.mapred.index.TableMap.map(TableMap.java:54)
       at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:50)
       at org.apache.hadoop.mapred.MapTask.run(MapTask.java:208)
at org.apache.hadoop.mapred.TaskTracker$Child.main(TaskTracker.java:2043)

org.apache.hadoop.util.DiskChecker$DiskErrorException: Could not find any valid local directory for task_200801101958_0001_m_000358_0/spill0.out at org.apache.hadoop.fs.LocalDirAllocator$AllocatorPerContext.getLocalPathForWrite(LocalDirAllocator.java:313) at org.apache.hadoop.fs.LocalDirAllocator.getLocalPathForWrite(LocalDirAllocator.java:124) at org.apache.hadoop.mapred.MapOutputFile.getSpillFileForWrite(MapOutputFile.java:88) at org.apache.hadoop.mapred.MapTask$MapOutputBuffer.sortAndSpillToDisk(MapTask.java:405) at org.apache.hadoop.mapred.MapTask$MapOutputBuffer.collect(MapTask.java:387)
       at com.worldlingo.hadoop.mapred.index.TableMap.map(TableMap.java:54)
       at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:50)
       at org.apache.hadoop.mapred.MapTask.run(MapTask.java:208)
at org.apache.hadoop.mapred.TaskTracker$Child.main(TaskTracker.java:2043)

org.apache.hadoop.util.DiskChecker$DiskErrorException: Could not find any valid local directory for task_200801101958_0001_m_000373_0/spill0.out at org.apache.hadoop.fs.LocalDirAllocator$AllocatorPerContext.getLocalPathForWrite(LocalDirAllocator.java:313) at org.apache.hadoop.fs.LocalDirAllocator.getLocalPathForWrite(LocalDirAllocator.java:124) at org.apache.hadoop.mapred.MapOutputFile.getSpillFileForWrite(MapOutputFile.java:88) at org.apache.hadoop.mapred.MapTask$MapOutputBuffer.sortAndSpillToDisk(MapTask.java:405) at org.apache.hadoop.mapred.MapTask$MapOutputBuffer.collect(MapTask.java:387)
       at com.worldlingo.hadoop.mapred.index.TableMap.map(TableMap.java:54)
       at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:50)
Error initializing task_200801101958_0001_m_000390_0: org.apache.hadoop.util.DiskChecker$DiskErrorException: Could not find any valid local directory for taskTracker/jobcache/job_200801101958_0001/job.xml at org.apache.hadoop.fs.LocalDirAllocator$AllocatorPerContext.getLocalPathForWrite(LocalDirAllocator.java:313) at org.apache.hadoop.fs.LocalDirAllocator.getLocalPathForWrite(LocalDirAllocator.java:124) at org.apache.hadoop.mapred.TaskTracker.localizeJob(TaskTracker.java:635) at org.apache.hadoop.mapred.TaskTracker.startNewTask(TaskTracker.java:1264) at org.apache.hadoop.mapred.TaskTracker.offerService(TaskTracker.java:920)
       at org.apache.hadoop.mapred.TaskTracker.run(TaskTracker.java:1300)
       at org.apache.hadoop.mapred.TaskTracker.main(TaskTracker.java:2163)

08/01/10 20:26:49 WARN mapred.JobClient: Error reading task outputhttp://lv1-xen-pdc-60.worldlingo.com:50060/tasklog?plaintext=true&taskid=task_200801101958_0001_m_000390_0&filter=stdout 08/01/10 20:26:49 WARN mapred.JobClient: Error reading task outputhttp://lv1-xen-pdc-60.worldlingo.com:50060/tasklog?plaintext=true&taskid=task_200801101958_0001_m_000390_0&filter=stderr org.apache.hadoop.util.DiskChecker$DiskErrorException: Could not find any valid local directory for task_200801101958_0001_m_000371_0/spill3.out at org.apache.hadoop.fs.LocalDirAllocator$AllocatorPerContext.getLocalPathForWrite(LocalDirAllocator.java:313) at org.apache.hadoop.fs.LocalDirAllocator.getLocalPathForWrite(LocalDirAllocator.java:124) at org.apache.hadoop.mapred.MapOutputFile.getSpillFileForWrite(MapOutputFile.java:88) at org.apache.hadoop.mapred.MapTask$MapOutputBuffer.sortAndSpillToDisk(MapTask.java:405) at org.apache.hadoop.mapred.MapTask$MapOutputBuffer.flush(MapTask.java:627)
       at org.apache.hadoop.mapred.MapTask.run(MapTask.java:209)
at org.apache.hadoop.mapred.TaskTracker$Child.main(TaskTracker.java:2043)

Shuffle Error: Exceeded MAX_FAILED_UNIQUE_FETCHES; bailing-out.
org.apache.hadoop.hbase.UnknownScannerException: org.apache.hadoop.hbase.UnknownScannerException: 3314387006816878921 at org.apache.hadoop.hbase.HRegionServer.close(HRegionServer.java:1494)
       at sun.reflect.GeneratedMethodAccessor9.invoke(Unknown Source)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
       at java.lang.reflect.Method.invoke(Method.java:585)
       at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:401)
       at org.apache.hadoop.ipc.Server$Handler.run(Server.java:892)

at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:39) at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27)
       at java.lang.reflect.Constructor.newInstance(Constructor.java:494)
at org.apache.hadoop.hbase.RemoteExceptionHandler.decodeRemoteException(RemoteExceptionHandler.java:82) at org.apache.hadoop.hbase.HTable$ClientScanner.close(HTable.java:1099) at com.worldlingo.hadoop.mapred.index.TableInputFormat$TableRecordReader.close(TableInputFormat.java:65) at org.apache.hadoop.mapred.MapTask$TrackedRecordReader.close(MapTask.java:155)
       at org.apache.hadoop.mapred.MapTask.run(MapTask.java:212)
at org.apache.hadoop.mapred.TaskTracker$Child.main(TaskTracker.java:2043) I am an absolute noobie at all of this. Could anyone help me understand what went wrong, and what I have to do to get around it?

Thanks,
Lars

---
Lars George, CTO
WorldLingo

Reply via email to