I tried one more script:
A = load 'hbase://query/SELECT * from test_table' using
org.apache.phoenix.pig.
PhoenixHBaseLoader('ZK');
grpd = GROUP A BY Rowkey_Prefix_String;
cnt = FOREACH grpd GENERATE group AS Rowkey_Prefix_String,COUNT(A);
DUMP cnt;
Rowkey_Prefix_String is prefix of my composite rowkey.
Again, i got same error:
ERROR org.apache.hadoop.hbase.io.HbaseObjectWritable - Error in readFields
java.io.IOException: Error in readFields
at
org.apache.hadoop.hbase.io.HbaseObjectWritable.readObject(HbaseObjectWritable.java:698)
at
org.apache.hadoop.hbase.io.HbaseObjectWritable.readObject(HbaseObjectWritable.java:596)
at
org.apache.hadoop.hbase.client.coprocessor.ExecResult.readFields(ExecResult.java:83)
at
org.apache.hadoop.hbase.io.HbaseObjectWritable.readObject(HbaseObjectWritable.java:692)
at
org.apache.hadoop.hbase.io.HbaseObjectWritable.readFields(HbaseObjectWritable.java:333)
at
org.apache.hadoop.hbase.ipc.SecureClient$SecureConnection.receiveResponse(SecureClient.java:383)
at
org.apache.hadoop.hbase.ipc.HBaseClient$Connection.run(HBaseClient.java:588)
Caused by: java.lang.NegativeArraySizeException: -1
at org.apache.hadoop.hbase.util.Bytes.readByteArray(Bytes.java:175)
at
org.apache.phoenix.schema.PColumnImpl.readFields(PColumnImpl.java:157)
at org.apache.phoenix.schema.PTableImpl.readFields(PTableImpl.java:721)
at
org.apache.phoenix.coprocessor.MetaDataProtocol$MetaDataMutationResult.readFields(MetaDataProtocol.java:161)
at
org.apache.hadoop.hbase.io.HbaseObjectWritable.readObject(HbaseObjectWritable.java:692)
... 6 more
I am able to query this table from sqlline.
Thanks,
Anil Gupta
On Tue, Jul 1, 2014 at 1:51 PM, anil gupta <[email protected]> wrote:
> Hi All,
>
> I am running the HBase0.94.15 and latest phoenix 3.1 nightly build. I have
> to use pig on phoenix views. When i run the job i get following error:
> ERROR org.apache.hadoop.hbase.io.HbaseObjectWritable - Error in readFields
> java.lang.NegativeArraySizeException: -1
> at org.apache.hadoop.hbase.util.Bytes.readByteArray(Bytes.java:175)
> at
> org.apache.phoenix.schema.PColumnImpl.readFields(PColumnImpl.java:157)
> at org.apache.phoenix.schema.PTableImpl.readFields(PTableImpl.java:721)
> at
> org.apache.phoenix.coprocessor.MetaDataProtocol$MetaDataMutationResult.readFields(MetaDataProtocol.java:161)
> at
> org.apache.hadoop.hbase.io.HbaseObjectWritable.readObject(HbaseObjectWritable.java:692)
> at
> org.apache.hadoop.hbase.io.HbaseObjectWritable.readObject(HbaseObjectWritable.java:596)
> at
> org.apache.hadoop.hbase.client.coprocessor.ExecResult.readFields(ExecResult.java:83)
> at
> org.apache.hadoop.hbase.io.HbaseObjectWritable.readObject(HbaseObjectWritable.java:692)
> at
> org.apache.hadoop.hbase.io.HbaseObjectWritable.readFields(HbaseObjectWritable.java:333)
> at
> org.apache.hadoop.hbase.ipc.SecureClient$SecureConnection.receiveResponse(SecureClient.java:383)
> at
> org.apache.hadoop.hbase.ipc.HBaseClient$Connection.run(HBaseClient.java:588)
>
> Here is my pig script:
> A = load 'hbase://query/SELECT * from test_table' using
> org.apache.phoenix.pig.PhoenixHBaseLoader('ZK');
> grpd = GROUP A BY UNSIGNED_DATE_COLUMN;
> cnt = FOREACH grpd GENERATE group AS UNSIGNED_DATE_COLUMN,COUNT(A);
> DUMP cnt;
>
> Can anyone tell me what the issue over here? I am suspecting maybe
> UNSIGNED_DATE type column is not supported in Pig integration?
>
>
> --
> Thanks & Regards,
> Anil Gupta
>
--
Thanks & Regards,
Anil Gupta