I have a table with a primary key that performs well, as well as 2 indexes, 
which I created like this:

CREATE INDEX <my_table_prefix>_indexed_meterkey_v2 on 
<my_table_prefix>_indexed_meterkey_immutable_v2 (meter_key)

(<my_table_prefix> and <my server> is just some obfuscation for the purposes of 
posting here)

We WERE  running Phoenix 4.6, which I had manually patched onto our HDP 2.2.8 
stack.

We recently upgraded to HDP 2.4.2, which includes Phoenix 4.4 (technically it's 
4.4.0.2.4.2.0)

I couldn't get the 4.6 stuff to work on this stack (at least not easily), and I 
really don't need it. I originally made the change to 4.6 to get the 
QueryServices and PhoenixRuntime packages, which I no longer use.

So I removed 4.6 and I'm running with 4.4 everywhere.

It mostly works, with the notable exception that now, when I run CSV bulk 
upload of data, it fails to insert into the index tables, as described in the 
stack track below.

So, MY QUESTIONs:

1) Is there a known solution to fix this?
Or
2) I can live without the indexes. Can someone explain to me how to remove the 
indexes/index tables without harming the base tables/key?

Thanks!


16/05/23 10:48:58 INFO zookeeper.RecoverableZooKeeper: Process 
identifier=hconnection-0x31e924f9 connecting to ZooKeeper 
ensemble=<server1>:2181,<server2>:2181,<server3>:2181
16/05/23 10:48:58 INFO zookeeper.ZooKeeper: Initiating client connection, 
connectString=<server1>:2181,<server2>:2181,<server3>:2181 
sessionTimeout=120000 watcher=hconnection-0x31e924f90x0, 
quorum=<server1>:2181,<server2>:2181,<server3>:2181, baseZNode=/hbase
16/05/23 10:48:58 INFO zookeeper.ClientCnxn: Opening socket connection to 
server <server2>/10.22.4.65:2181. Will not attempt to authenticate using SASL 
(unknown error)
16/05/23 10:48:58 INFO zookeeper.ClientCnxn: Socket connection established to 
<server2>/10.22.4.65:2181, initiating session
16/05/23 10:48:58 INFO zookeeper.ClientCnxn: Session establishment complete on 
server <server2>/10.22.4.65:2181, sessionid = 0x254c9555ca7020b, negotiated 
timeout = 40000
16/05/23 10:48:58 ERROR mapreduce.CsvBulkLoadTool: Import job on 
table=<my_table_prefix>_INDEXED_FLDREPID_V2 failed due to exception.
org.apache.hadoop.hbase.TableNotFoundException: 
<my_table_prefix>_INDEXED_FLDREPID_V2
                at 
org.apache.hadoop.hbase.client.HTable.getTableDescriptor(HTable.java:597)
                at 
org.apache.hadoop.hbase.mapreduce.HFileOutputFormat.configureIncrementalLoad(HFileOutputFormat.java:91)
                at 
org.apache.phoenix.mapreduce.CsvBulkLoadTool$TableLoader.call(CsvBulkLoadTool.java:512)
                at 
org.apache.phoenix.mapreduce.CsvBulkLoadTool$TableLoader.call(CsvBulkLoadTool.java:473)
                at java.util.concurrent.FutureTask.run(FutureTask.java:262)
                at 
org.apache.phoenix.job.JobManager$InstrumentedJobFutureTask.run(JobManager.java:172)
                at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
                at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
                at java.lang.Thread.run(Thread.java:745)
16/05/23 10:48:58 ERROR mapreduce.CsvBulkLoadTool: Import job on 
table=<my_table_prefix>_INDEXED_METERKEY_V2 failed due to exception.
org.apache.hadoop.hbase.TableNotFoundException: 
<my_table_prefix>_HASHED_INDEXED_METERKEY_V2
                at 
org.apache.hadoop.hbase.client.HTable.getTableDescriptor(HTable.java:597)
                at 
org.apache.hadoop.hbase.mapreduce.HFileOutputFormat.configureIncrementalLoad(HFileOutputFormat.java:91)
                at 
org.apache.phoenix.mapreduce.CsvBulkLoadTool$TableLoader.call(CsvBulkLoadTool.java:512)
                at 
org.apache.phoenix.mapreduce.CsvBulkLoadTool$TableLoader.call(CsvBulkLoadTool.java:473)
                at java.util.concurrent.FutureTask.run(FutureTask.java:262)
                at 
org.apache.phoenix.job.JobManager$InstrumentedJobFutureTask.run(JobManager.java:172)
                at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
                at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
                at java.lang.Thread.run(Thread.java:745)
16/05/23 10:48:58 INFO mapreduce.HFileOutputFormat2: Looking up current regions 
for table <my_table_prefix>_INDEXED_METERKEY_IMMUTABLE_V2
16/05/23 10:48:58 WARN hbase.HBaseConfiguration: Config option 
"hbase.regionserver.lease.period" is deprecated. Instead, use 
"hbase.client.scanner.timeout.period"
16/05/23 10:48:58 INFO mapreduce.HFileOutputFormat2: Configuring 11 reduce 
partitions to match current region count
16/05/23 10:48:58 INFO mapreduce.HFileOutputFormat2: Writing partition 
information to 
/user/bcampbell/hbase-staging/partitions_23aea473-3b08-4827-bfff-a8dd29356089
16/05/23 10:48:58 INFO zlib.ZlibFactory: Successfully loaded & initialized 
native-zlib library
16/05/23 10:48:58 INFO compress.CodecPool: Got brand-new compressor [.deflate]
16/05/23 10:48:58 INFO mapreduce.HFileOutputFormat2: Incremental table 
<my_table_prefix>_INDEXED_METERKEY_IMMUTABLE_V2 output configured.
16/05/23 10:48:58 INFO mapreduce.CsvBulkLoadTool: Running MapReduce import job 
from /user/hdfs/HBASE_STAGING/edd_v12_staging_1463270400-1464014846 to 
/tmp/c11fa527-84ba-492b-ac36-562810c43208/<my_table_prefix>_INDEXED_METERKEY_IMMUTABLE_V2
16/05/23 10:48:58 INFO impl.TimelineClientImpl: Timeline service address: 
http://<server2>:8188/ws/v1/timeline/
16/05/23 10:48:58 INFO client.RMProxy: Connecting to ResourceManager at 
<server2>/10.22.4.65:8050
16/05/23 10:48:59 INFO input.FileInputFormat: Total input paths to process : 1
16/05/23 10:48:59 INFO mapreduce.JobSubmitter: number of splits:8
16/05/23 10:48:59 INFO Configuration.deprecation: io.bytes.per.checksum is 
deprecated. Instead, use dfs.bytes-per-checksum
16/05/23 10:48:59 INFO mapreduce.JobSubmitter: Submitting tokens for job: 
job_1463765948801_0291


Reply via email to