What version of Pig are you running and what errors are you seeing on the task trackers?
On Tue, Apr 26, 2011 at 4:46 AM, byambajargal <byambaa.0...@gmail.com> wrote: > Hello ... > I have a question for you > > I am doing a pig job as following that read from hdfs simply to store hbase > when i start the job first part works fine and second part was failure. > Could you give me a direction how to move data from hdfs to Hbase > > > A = load '/passwd' using PigStorage(':');B = foreach A generate $0 as id, > $2 as value;dump B; > store B into 'table2' using > org.apache.pig.backend.hadoop.hbase.HBaseStorage( 'cf:a' , '-loadKey'); > > thank you for your help > > Byambajargal > > > > On 4/25/11 18:26, Dmitriy Ryaboy wrote: >> >> The first element of the relation you store must be the row key. You >> aren't >> loading the row key, so load> store isn't working. >> Try >> my_data = LOAD 'hbase://table1' using >> org.apache.pig.backend.hadoop.hbase.HBaseStorage('cf:1', '-loadKey') ; >> >> On Mon, Apr 25, 2011 at 5:32 AM, >> byambajargal<byambaa.0...@gmail.com>wrote: >> >>> Hello guys >>> >>> I am running cloudere distribution cdh3u0 on my cluster with Pig and >>> Hbase. >>> i can read data from hbase using the following pig query: >>> >>> my_data = LOAD 'hbase://table1' using >>> org.apache.pig.backend.hadoop.hbase.HBaseStorage('cf:1') ;dump my_data >>> >>> but when i try to store data into hbase as same way the job was failure. >>> >>> store my_data into 'hbase://table2' using >>> org.apache.pig.backend.hadoop.hbase.HBaseStorage('cf:1'); >>> >>> the table1 and the table2 has same structure and same column. >>> >>> >>> the table i have: >>> >>> hbase(main):029:0* scan 'table1' >>> ROW COLUMN+CELL >>> row1 column=cf:1, timestamp=1303731834050, value=value1 >>> row2 column=cf:1, timestamp=1303731849901, value=value2 >>> row3 column=cf:1, timestamp=1303731858637, value=value3 >>> 3 row(s) in 0.0470 seconds >>> >>> >>> thanks >>> >>> Byambajargal >>> >>> >>> > >