How did you export the table? Using HBase's Export? If so, in 0.94+ (HBASE-5440, should not be hard to backport of 0.92) you can use Import to create HFiles for later use with bulk import (LoadIncrementalHFiles).
Creating the HFiles takes far less time, and loading them into HBase in nearly instantaneous (provided that they were created recent enough, so they do not have to be re-split, because some of the tables region split). -- Lars ________________________________ From: Rita <[email protected]> To: [email protected] Sent: Thursday, March 29, 2012 7:57 PM Subject: importing a large table Hello, I am importing a 40+ billion row table which I exported several months ago. The data size is close to 18TB on hdfs (3x replication). My problem is when I try to import it with mapreduce it takes a few days -- which is ok -- however when the job fails to whatever reason, I have to restart everything. Is it possible to import the table in chunks like, import 1/3, 2/3, and then finally 3/3 of the table? Btw, the jobs creates close to 150k mapper jobs, thats a problem waiting to happen :-) -- --- Get your facts first, then you can distort them as you please.--
