Hi Ralph, You may want to check this problem against the latest release of Phoenix, coz we just incorporated a fix for a similar issue in our 3.2.1 RC1 and 4.2.1 RC1.
Thanks, Maryann On Mon, Nov 17, 2014 at 6:32 PM, Maryann Xue <[email protected]> wrote: > Hi Ralph, > > I think this is a known issue reported as PHOENIX-1011 ( > https://issues.apache.org/jira/browse/PHOENIX-1011). We are still looking > at it. Will give you an update once it is solved. > > Thanks a lot for the very detailed information, Ralph! > > > Thanks, > Maryann > > On Mon, Nov 17, 2014 at 12:24 PM, Perko, Ralph J <[email protected]> > wrote: > >> Hi, while importing data using the CsvBulkLoadTool I’ve run into an >> issue trying to query the data using sqlline.py. The bulk load tool was >> successful. There were no errors. However when I attempt to query the >> data I get some exceptions: >> >> >> >> java.lang.RuntimeException: >> org.apache.phoenix.exception.PhoenixIOException >> >> at sqlline.SqlLine$IncrementalRows.hasNext(SqlLine.java:2440) >> >> >> >> followed by many GlobalMemoryManger errors: >> >> >> >> WARN memory.GlobalMemoryManager: Orphaned chunk of xxxx bytes found >> during finalize >> >> >> >> Not all queries, but most, produce this error and it seems related to the >> existence of a secondary index table: >> >> >> >> select * from TABLE limit 10; --ERROR – index not used >> >> select <un-indexed field> from TABLE limit 10 -- ERROR >> >> >> >> If I run a query on an INTEGER column with a secondary index I do not get >> this error: >> >> >> >> select distinct(fieldx) from TABLE limit 10; -- SUCCESS! >> >> >> >> However, a similar query on an indexed VARCHAR field produces a timeout >> error: >> >> java.lang.RuntimeException: … PhoenixIOException: Failed after retry of >> OutOfOrderScannerNextException: was there a rpc timeout? >> >> at sqlline.SqlLine$IncrementalRows.hasNext(SqlLine.java:2440) >> >> >> >> select count(*) … times out as well >> >> >> >> Details: >> >> Total records imported: 7.2B >> >> Cluster size: 30 nodes >> >> Splits: 40 (salted) >> >> >> >> Phoenix version: 4.2.0 >> >> HBase version: 0.98 >> >> HDP distro 2.1.5 >> >> >> >> I can scan the data with no errors from hbase shell >> >> >> >> Basic Phoenix table def: >> >> >> >> CREATE TABLE IF NOT EXISTS >> >> t1_csv_data >> >> ( >> >> timestamp BIGINT NOT NULL, >> >> location VARCHAR NOT NULL, >> >> fileid VARCHAR NOT NULL, >> >> recnum INTEGER NOT NULL, >> >> field5 VARCHAR, >> >> ... >> >> field45 VARCHAR, >> >> CONSTRAINT pkey PRIMARY KEY (timestamp, >> >> location, fileid,recnum) >> >> ) >> >> IMMUTABLE_ROWS=true,COMPRESSION='SNAPPY',SALT_BUCKETS=40, >> SPLIT_POLICY=’org.apache.hadoop.hbase.regionserver.ConstantSizeRegionSplitPolicy’; >> >> >> >> -- indexes >> >> CREATE INDEX t1_csv_data_f1_idx ON t1_csv_data(somefield1) >> COMPRESSION='SNAPPY', >> SPLIT_POLICY=’org.apache.hadoop.hbase.regionserver.ConstantSizeRegionSplitPolicy’;; >> >> CREATE INDEX t1_csv_data_f2_idx ON t1_csv_data(somefield2) >> COMPRESSION='SNAPPY', >> SPLIT_POLICY=’org.apache.hadoop.hbase.regionserver.ConstantSizeRegionSplitPolicy’;; >> >> CREATE INDEX t1_csv_data_f3_idx ON t1_csv_data(somefield3) >> COMPRESSION='SNAPPY', >> SPLIT_POLICY=’org.apache.hadoop.hbase.regionserver.ConstantSizeRegionSplitPolicy’;; >> >> >> >> Thanks for your help, >> >> Ralph >> >> >> > > > > -- > Thanks, > Maryann > -- Thanks, Maryann
