Hello Guys,

Hope everybody keeping good.

I have few questions -

Tested ON - Phoenix4.1 , HBase0.98.1 , 7 RS heap Size - 8 GB

1) PIG gives error whenever I am trying to LOAD data from Salted Phoenix
table

A = LOAD 'hbase://table/RAW_LOG' USING
org.apache.phoenix.pig.PhoenixHBaseLoader('pari');

2) STATS not working through SQLLine
 UPDATE STATISTICS my_table ;
Error: ERROR 601 (42P00): Syntax error. Encountered "UPDATE" at line 1,
column 1. (state=42P00,code=601)

http://phoenix.apache.org/update_statistics.html

3) We have RAW table with ~25M rows, we want to load aggregated data to new
table. Everything works if we use upto two aggregate functions. But fails
with
Error: org.apache.phoenix.exception.PhoenixIOException:
org.apache.hadoop.hbase.DoNotRetryIOException:
G_DEV_YDSP.RAW_LOG,\x0F\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00,1415059496937.22f088da9f0e53f6fc0ad1275e901ceb.:
null
        at
org.apache.phoenix.util.ServerUtil.createIOException(ServerUtil.java:77)
        at
org.apache.phoenix.util.ServerUtil.throwIOException(ServerUtil.java:45)
        at
org.apache.phoenix.coprocessor.BaseScannerRegionObserver.postScannerOpen(BaseScannerRegionObserver.java:152)
        at
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.postScannerOpen(RegionCoprocessorHost.java:1663)
        at
org.apache.hadoop.hbase.regionserver.HRegionServer.scan(HRegionServer.java:3071)
        at
org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:29497)
        at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2012)
        at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:98)
        at
org.apache.hadoop.hbase.ipc.SimpleRpcScheduler.consumerLoop(SimpleRpcScheduler.java:160)
        at
org.apache.hadoop.hbase.ipc.SimpleRpcScheduler.access$000(SimpleRpcScheduler.java:38)
        at
org.apache.hadoop.hbase.ipc.SimpleRpcScheduler$1.run(SimpleRpcScheduler.java:110)
        at java.lang.Thread.run(Thread.java:744)
Caused by: java.lang.ArrayIndexOutOfBoundsException
        at org.apache.hadoop.hbase.util.Bytes.putBytes(Bytes.java:290)
        at
org.apache.hadoop.hbase.KeyValue.createByteArray(KeyValue.java:1031)
        at org.apache.hadoop.hbase.KeyValue.<init>(KeyValue.java:639)
        at org.apache.hadoop.hbase.KeyValue.<init>(KeyValue.java:567)
        at
org.apache.phoenix.util.KeyValueUtil.newKeyValue(KeyValueUtil.java:63)
        at
org.apache.phoenix.cache.aggcache.SpillManager.getAggregators(SpillManager.java:200)
        at
org.apache.phoenix.cache.aggcache.SpillManager.loadEntry(SpillManager.java:273)
        at
org.apache.phoenix.cache.aggcache.SpillableGroupByCache.cache(SpillableGroupByCache.java:231)
        at org.apache.phoenix.coprocessor.GroupedAggregateRegionObserver
.scanUnordered(GroupedAggregateRegionObserver.java:432)
        at
org.apache.phoenix.coprocessor.GroupedAggregateRegionObserver.doPostScannerOpen(GroupedAggregateRegionObserver.java:161)
        at
org.apache.phoenix.coprocessor.BaseScannerRegionObserver.postScannerOpen(BaseScannerRegionObserver.java:134)

-- 
Cheers,
Pari

Reply via email to