My command:
 hadoop jar $HBASE_HOME/hbase-0.90.4-cdh3u2.jar importtsv 
-Dimporttsv.separator=, -Dimporttsv.bulk.output=/tmp/output 
-Dimporttsv.columns=HBASE_ROW_KEY,e:a,e:b,e:c t1 /tmp/1


Usage: importtsv -Dimporttsv.columns=a,b,c <tablename> <inputdir>

Imports the given input directory of TSV data into the specified table.

The column names of the TSV data must be specified using the 
-Dimporttsv.columns option. This option takes the form of comma-separated 
column names, where each column name is either a simple column family, or a 
columnfamily:qualifier. The special column name HBASE_ROW_KEY is used to 
designate that this column should be used as the row key for each imported 
record. You must specify exactly one column to be the row key, and you must 
specify a column name for every column that exists in the input data.

In order to prepare data for a bulk data load, pass the option:
  -Dimporttsv.bulk.output=/path/for/output
  Note: if you do not use this option, then the target table must already exist 
in HBase            --look this line.

Other options that may be specified with -D include:
  -Dimporttsv.skip.bad.lines=false - fail if encountering an invalid line
  '-Dimporttsv.separator=|' - eg separate on pipes instead of tabs


-----邮件原件-----
发件人: [email protected] [mailto:[email protected]] 代表 Stack
发送时间: 2011年12月9日 14:24
收件人: [email protected]
主题: Re: 答复: TableNotFoundException: Cannot find row in .META. for table

On Thu, Dec 8, 2011 at 9:32 PM, Dou Xiaofeng <[email protected]> wrote:
> The table t1 not exist.
> If I create it by hbase client manually, the importtsv does not throw error, 
> but I assign the bulk.output in the command, it should be not to create the 
> table.
>

Sorry, I don't follow the last bit of the sentence above where you say
'... but I assign the bulk.output in the command....'

St.Ack

Reply via email to