I have a table with a primary key that performs well, as well as 2 indexes,
which I created like this:
CREATE INDEX <my_table_prefix>_indexed_meterkey_v2 on
<my_table_prefix>_indexed_meterkey_immutable_v2 (meter_key)
(<my_table_prefix> and <my server> is just some obfuscation for the purposes of
posting here)
We WERE running Phoenix 4.6, which I had manually patched onto our HDP 2.2.8
stack.
We recently upgraded to HDP 2.4.2, which includes Phoenix 4.4 (technically it's
4.4.0.2.4.2.0)
I couldn't get the 4.6 stuff to work on this stack (at least not easily), and I
really don't need it. I originally made the change to 4.6 to get the
QueryServices and PhoenixRuntime packages, which I no longer use.
So I removed 4.6 and I'm running with 4.4 everywhere.
It mostly works, with the notable exception that now, when I run CSV bulk
upload of data, it fails to insert into the index tables, as described in the
stack track below.
So, MY QUESTIONs:
1) Is there a known solution to fix this?
Or
2) I can live without the indexes. Can someone explain to me how to remove the
indexes/index tables without harming the base tables/key?
Thanks!
16/05/23 10:48:58 INFO zookeeper.RecoverableZooKeeper: Process
identifier=hconnection-0x31e924f9 connecting to ZooKeeper
ensemble=<server1>:2181,<server2>:2181,<server3>:2181
16/05/23 10:48:58 INFO zookeeper.ZooKeeper: Initiating client connection,
connectString=<server1>:2181,<server2>:2181,<server3>:2181
sessionTimeout=120000 watcher=hconnection-0x31e924f90x0,
quorum=<server1>:2181,<server2>:2181,<server3>:2181, baseZNode=/hbase
16/05/23 10:48:58 INFO zookeeper.ClientCnxn: Opening socket connection to
server <server2>/10.22.4.65:2181. Will not attempt to authenticate using SASL
(unknown error)
16/05/23 10:48:58 INFO zookeeper.ClientCnxn: Socket connection established to
<server2>/10.22.4.65:2181, initiating session
16/05/23 10:48:58 INFO zookeeper.ClientCnxn: Session establishment complete on
server <server2>/10.22.4.65:2181, sessionid = 0x254c9555ca7020b, negotiated
timeout = 40000
16/05/23 10:48:58 ERROR mapreduce.CsvBulkLoadTool: Import job on
table=<my_table_prefix>_INDEXED_FLDREPID_V2 failed due to exception.
org.apache.hadoop.hbase.TableNotFoundException:
<my_table_prefix>_INDEXED_FLDREPID_V2
at
org.apache.hadoop.hbase.client.HTable.getTableDescriptor(HTable.java:597)
at
org.apache.hadoop.hbase.mapreduce.HFileOutputFormat.configureIncrementalLoad(HFileOutputFormat.java:91)
at
org.apache.phoenix.mapreduce.CsvBulkLoadTool$TableLoader.call(CsvBulkLoadTool.java:512)
at
org.apache.phoenix.mapreduce.CsvBulkLoadTool$TableLoader.call(CsvBulkLoadTool.java:473)
at java.util.concurrent.FutureTask.run(FutureTask.java:262)
at
org.apache.phoenix.job.JobManager$InstrumentedJobFutureTask.run(JobManager.java:172)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)
16/05/23 10:48:58 ERROR mapreduce.CsvBulkLoadTool: Import job on
table=<my_table_prefix>_INDEXED_METERKEY_V2 failed due to exception.
org.apache.hadoop.hbase.TableNotFoundException:
<my_table_prefix>_HASHED_INDEXED_METERKEY_V2
at
org.apache.hadoop.hbase.client.HTable.getTableDescriptor(HTable.java:597)
at
org.apache.hadoop.hbase.mapreduce.HFileOutputFormat.configureIncrementalLoad(HFileOutputFormat.java:91)
at
org.apache.phoenix.mapreduce.CsvBulkLoadTool$TableLoader.call(CsvBulkLoadTool.java:512)
at
org.apache.phoenix.mapreduce.CsvBulkLoadTool$TableLoader.call(CsvBulkLoadTool.java:473)
at java.util.concurrent.FutureTask.run(FutureTask.java:262)
at
org.apache.phoenix.job.JobManager$InstrumentedJobFutureTask.run(JobManager.java:172)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)
16/05/23 10:48:58 INFO mapreduce.HFileOutputFormat2: Looking up current regions
for table <my_table_prefix>_INDEXED_METERKEY_IMMUTABLE_V2
16/05/23 10:48:58 WARN hbase.HBaseConfiguration: Config option
"hbase.regionserver.lease.period" is deprecated. Instead, use
"hbase.client.scanner.timeout.period"
16/05/23 10:48:58 INFO mapreduce.HFileOutputFormat2: Configuring 11 reduce
partitions to match current region count
16/05/23 10:48:58 INFO mapreduce.HFileOutputFormat2: Writing partition
information to
/user/bcampbell/hbase-staging/partitions_23aea473-3b08-4827-bfff-a8dd29356089
16/05/23 10:48:58 INFO zlib.ZlibFactory: Successfully loaded & initialized
native-zlib library
16/05/23 10:48:58 INFO compress.CodecPool: Got brand-new compressor [.deflate]
16/05/23 10:48:58 INFO mapreduce.HFileOutputFormat2: Incremental table
<my_table_prefix>_INDEXED_METERKEY_IMMUTABLE_V2 output configured.
16/05/23 10:48:58 INFO mapreduce.CsvBulkLoadTool: Running MapReduce import job
from /user/hdfs/HBASE_STAGING/edd_v12_staging_1463270400-1464014846 to
/tmp/c11fa527-84ba-492b-ac36-562810c43208/<my_table_prefix>_INDEXED_METERKEY_IMMUTABLE_V2
16/05/23 10:48:58 INFO impl.TimelineClientImpl: Timeline service address:
http://<server2>:8188/ws/v1/timeline/
16/05/23 10:48:58 INFO client.RMProxy: Connecting to ResourceManager at
<server2>/10.22.4.65:8050
16/05/23 10:48:59 INFO input.FileInputFormat: Total input paths to process : 1
16/05/23 10:48:59 INFO mapreduce.JobSubmitter: number of splits:8
16/05/23 10:48:59 INFO Configuration.deprecation: io.bytes.per.checksum is
deprecated. Instead, use dfs.bytes-per-checksum
16/05/23 10:48:59 INFO mapreduce.JobSubmitter: Submitting tokens for job:
job_1463765948801_0291