Hi Ravi, Here is sample DDL for table: CREATE view test_table( Rowkey_Prefix_String VARCHAR not null, rev_time BIGINT not null, ora_id BIGINT not null, src_db_id SMALLINT, t.big_decimal_val DECIMAL, UNSIGNED_DATE_COLUMN UNSIGNED_DATE, t.OID VARCHAR CONSTRAINT pk PRIMARY KEY(Rowkey_Prefix_String ,rev_time,ora_id));
On Tue, Jul 1, 2014 at 2:48 PM, Ravi Kiran <[email protected]> wrote: > Hi Anil, > Can you please share the table ddl to help me write tests to see where > the issue is. > > Regards > Ravi > > > On Tue, Jul 1, 2014 at 1:58 PM, anil gupta <[email protected]> wrote: > >> I tried one more script: >> A = load 'hbase://query/SELECT * from test_table' using >> org.apache.phoenix.pig. >> PhoenixHBaseLoader('ZK'); >> grpd = GROUP A BY Rowkey_Prefix_String; >> cnt = FOREACH grpd GENERATE group AS Rowkey_Prefix_String,COUNT(A); >> DUMP cnt; >> >> Rowkey_Prefix_String is prefix of my composite rowkey. >> >> Again, i got same error: >> ERROR org.apache.hadoop.hbase.io.HbaseObjectWritable - Error in readFields >> java.io.IOException: Error in readFields >> at >> >> org.apache.hadoop.hbase.io.HbaseObjectWritable.readObject(HbaseObjectWritable.java:698) >> at >> >> org.apache.hadoop.hbase.io.HbaseObjectWritable.readObject(HbaseObjectWritable.java:596) >> at >> >> org.apache.hadoop.hbase.client.coprocessor.ExecResult.readFields(ExecResult.java:83) >> at >> >> org.apache.hadoop.hbase.io.HbaseObjectWritable.readObject(HbaseObjectWritable.java:692) >> at >> >> org.apache.hadoop.hbase.io.HbaseObjectWritable.readFields(HbaseObjectWritable.java:333) >> at >> >> org.apache.hadoop.hbase.ipc.SecureClient$SecureConnection.receiveResponse(SecureClient.java:383) >> at >> >> org.apache.hadoop.hbase.ipc.HBaseClient$Connection.run(HBaseClient.java:588) >> Caused by: java.lang.NegativeArraySizeException: -1 >> at org.apache.hadoop.hbase.util.Bytes.readByteArray(Bytes.java:175) >> at >> org.apache.phoenix.schema.PColumnImpl.readFields(PColumnImpl.java:157) >> at >> org.apache.phoenix.schema.PTableImpl.readFields(PTableImpl.java:721) >> at >> >> org.apache.phoenix.coprocessor.MetaDataProtocol$MetaDataMutationResult.readFields(MetaDataProtocol.java:161) >> at >> >> org.apache.hadoop.hbase.io.HbaseObjectWritable.readObject(HbaseObjectWritable.java:692) >> ... 6 more >> >> I am able to query this table from sqlline. >> >> Thanks, >> Anil Gupta >> >> >> >> >> On Tue, Jul 1, 2014 at 1:51 PM, anil gupta <[email protected]> wrote: >> >> > Hi All, >> > >> > I am running the HBase0.94.15 and latest phoenix 3.1 nightly build. I >> have >> > to use pig on phoenix views. When i run the job i get following error: >> > ERROR org.apache.hadoop.hbase.io.HbaseObjectWritable - Error in >> readFields >> > java.lang.NegativeArraySizeException: -1 >> > at org.apache.hadoop.hbase.util.Bytes.readByteArray(Bytes.java:175) >> > at >> > org.apache.phoenix.schema.PColumnImpl.readFields(PColumnImpl.java:157) >> > at >> org.apache.phoenix.schema.PTableImpl.readFields(PTableImpl.java:721) >> > at >> > >> org.apache.phoenix.coprocessor.MetaDataProtocol$MetaDataMutationResult.readFields(MetaDataProtocol.java:161) >> > at >> > >> org.apache.hadoop.hbase.io.HbaseObjectWritable.readObject(HbaseObjectWritable.java:692) >> > at >> > >> org.apache.hadoop.hbase.io.HbaseObjectWritable.readObject(HbaseObjectWritable.java:596) >> > at >> > >> org.apache.hadoop.hbase.client.coprocessor.ExecResult.readFields(ExecResult.java:83) >> > at >> > >> org.apache.hadoop.hbase.io.HbaseObjectWritable.readObject(HbaseObjectWritable.java:692) >> > at >> > >> org.apache.hadoop.hbase.io.HbaseObjectWritable.readFields(HbaseObjectWritable.java:333) >> > at >> > >> org.apache.hadoop.hbase.ipc.SecureClient$SecureConnection.receiveResponse(SecureClient.java:383) >> > at >> > >> org.apache.hadoop.hbase.ipc.HBaseClient$Connection.run(HBaseClient.java:588) >> > >> > Here is my pig script: >> > A = load 'hbase://query/SELECT * from test_table' using >> > org.apache.phoenix.pig.PhoenixHBaseLoader('ZK'); >> > grpd = GROUP A BY UNSIGNED_DATE_COLUMN; >> > cnt = FOREACH grpd GENERATE group AS UNSIGNED_DATE_COLUMN,COUNT(A); >> > DUMP cnt; >> > >> > Can anyone tell me what the issue over here? I am suspecting maybe >> > UNSIGNED_DATE type column is not supported in Pig integration? >> > >> > >> > -- >> > Thanks & Regards, >> > Anil Gupta >> > >> >> >> >> -- >> Thanks & Regards, >> Anil Gupta >> > > -- Thanks & Regards, Anil Gupta
