That is a mistake. Should be corrected.!

The way you are using is right.
The ID (first column) will be the hbase's rowkey. and the other's will get
into the columns you mention in the arg of HBaseStorage.


On Fri, Apr 20, 2012 at 6:23 PM, Royston Sellman <
[email protected]> wrote:

> OK, I'll ping the HBase folks...
>
> Meanwhile, are the HBaseStorage docs correct? The example shows the STORE
> command having 'USING' and 'AS' clauses, but 'AS' gives a parse error. 'AS'
> is valid in LOADs though.
>
> Cheers,
> Royston
>
>
> -----Original Message-----
> From: Dmitriy Ryaboy [mailto:[email protected]]
> Sent: 20 April 2012 00:03
> To: [email protected]
> Subject: Re: HBaseStorage not working
>
> Nothing significant changed in Pig trunk, so I am guessing HBase changed
> something; you are more likely to get help from them (they should at least
> be able to point at APIs that changed and are likely to cause this sort of
> thing).
>
> You might also want to check if any of the started MR jobs have anything
> interesting in their task logs.
>
> D
>
> On Thu, Apr 19, 2012 at 1:41 PM, Royston Sellman
> <[email protected]> wrote:
> > Does HBaseStorage work with HBase 0.95?
> >
> >
> >
> > This code was working with HBase 0.92 and Pig 0.9 but fails on HBase
> > 0.95 and Pig 0.11 (built from source):
> >
> >
> >
> > register /opt/hbase/hbase-trunk/hbase-0.95-SNAPSHOT.jar
> >
> > register /opt/zookeeper/zookeeper-3.4.3/zookeeper-3.4.3.jar
> >
> >
> >
> >
> >
> > tbl1 = LOAD 'input/sse.tbl1.HEADERLESS.csv' USING PigStorage( ',' ) AS
> > (
> >
> >      ID:chararray,
> >
> >      hp:chararray,
> >
> >      pf:chararray,
> >
> >      gz:chararray,
> >
> >      hid:chararray,
> >
> >      hst:chararray,
> >
> >      mgz:chararray,
> >
> >      gg:chararray,
> >
> >      epc:chararray );
> >
> >
> >
> > STORE tbl1 INTO 'hbase://sse.tbl1'
> >
> > USING org.apache.pig.backend.hadoop.hbase.HBaseStorage('edrp:hp
> > edrp:pf edrp:gz edrp:hid edrp:hst edrp:mgz edrp:gg edrp:epc');
> >
> >
> >
> > The job output (using either Grunt or PigServer makes no difference)
> > shows the family:descriptors being added by HBaseStorage then starts
> > up the MR job which (after a long pause) reports:
> >
> > ------------
> >
> > Input(s):
> >
> > Failed to read data from
> > "hdfs://namenode:8020/user/hadoop1/input/sse.tbl1.HEADERLESS.csv"
> >
> >
> >
> > Output(s):
> >
> > Failed to produce result in "hbase://sse.tbl1"
> >
> >
> >
> >
> >
> > INFO mapReduceLayer.MapReduceLauncher: Failed!
> >
> > INFO hbase.HBaseStorage: Adding family:descriptor filters with values
> > edrp:hp
> >
> > INFO hbase.HBaseStorage: Adding family:descriptor filters with values
> > edrp:pf
> >
> > INFO hbase.HBaseStorage: Adding family:descriptor filters with values
> > edrp:gz
> >
> > INFO hbase.HBaseStorage: Adding family:descriptor filters with values
> > edrp:hid
> >
> > INFO hbase.HBaseStorage: Adding family:descriptor filters with values
> > edrp:hst
> >
> > INFO hbase.HBaseStorage: Adding family:descriptor filters with values
> > edrp:mgz
> >
> > INFO hbase.HBaseStorage: Adding family:descriptor filters with values
> > edrp:gg
> >
> > INFO hbase.HBaseStorage: Adding family:descriptor filters with values
> > edrp:epc
> >
> > ------------
> >
> >
> >
> > The "Failed to read" is misleading I think because dump tbl1; in place
> > of the store works fine.
> >
> >
> >
> > I get nothing in the HBase logs and nothing in the Pig log.
> >
> >
> >
> > HBase works fine from the shell and can read and write to the table.
> > Pig works fine in and out of HDFS on CSVs.
> >
> >
> >
> > Any ideas?
> >
> >
> >
> > Royston
> >
> >
> >
>
>


-- 
Thanks and Regards,
Rajgopal Vaithiyanathan.

Reply via email to