Hello, Josh. Thank you for your response! I beg your pardon, our version is 5.1.0
There is what I see in PSQ logs with DEBUG enabled. 2020-01-30 13:20:57,483 DEBUG org.apache.phoenix.query.ConnectionQueryServicesImpl: System mutex table already appears to exist, not creating it 2020-01-30 13:20:57,521 DEBUG org.apache.phoenix.query.ConnectionQueryServicesImpl: Found quorum: hbase-data-upgrade001-stg.foo.bar:2181,hbase-data-upgrade002-stg.foo.bar:2181,hbase-master-upgrade001-stg.foo.bar:2181:/hbase 2020-01-30 13:20:57,557 DEBUG org.apache.phoenix.query.ConnectionQueryServicesImpl: 8406@hbase-master-upgrade001-stg acquired mutex for tenantId : null schemaName : SYSTEM tableName : CATALOG columnName : null familyName : null 2020-01-30 13:20:57,557 DEBUG org.apache.phoenix.query.ConnectionQueryServicesImpl: Acquired lock in SYSMUTEX table for migrating SYSTEM tables to SYSTEM namespace and/or upgrading SYSTEM:CATALOG 2020-01-30 13:20:57,559 DEBUG org.apache.phoenix.query.ConnectionQueryServicesImpl: Migrated SYSTEM tables to SYSTEM namespace 2020-01-30 13:20:57,559 INFO org.apache.phoenix.util.UpgradeUtil: Upgrading metadata to add parent links for indexes on views 2020-01-30 13:20:57,560 DEBUG org.apache.phoenix.jdbc.PhoenixStatement: {CurrentSCN=9223372036854775807} Execute query: SELECT TENANT_ID, TABLE_SCHEM, TABLE_NAME, COLUMN_FAMILY FROM SYSTEM.CATALOG WHERE LINK_TYPE = 1 2020-01-30 13:20:57,562 DEBUG org.apache.phoenix.execute.BaseQueryPlan: {CurrentSCN=9223372036854775807} Scan ready for iteration: {"loadColumnFamiliesOnDemand":true,"filter":"LINK_TYPE = 1","startRow":"","stopRow":"","batch":-1,"cacheBlocks":true,"totalColumns":0,"maxResultSize":-1,"families":{},"caching":2147483647,"maxVersions":1,"timeRange":[0,9223372036854775807]} 2020-01-30 13:20:57,562 DEBUG org.apache.phoenix.execute.BaseQueryPlan: {CurrentSCN=9223372036854775807} Iterator ready: org.apache.phoenix.iterate.RoundRobinResultIterator@27ba13aa 2020-01-30 13:20:57,562 DEBUG org.apache.phoenix.jdbc.PhoenixStatement: {CurrentSCN=9223372036854775807} Explain plan: CLIENT 1-CHUNK PARALLEL 1-WAY ROUND ROBIN FULL SCAN OVER SYSTEM:CATALOG SERVER FILTER BY LINK_TYPE = 1 2020-01-30 13:20:57,563 DEBUG org.apache.phoenix.iterate.BaseResultIterators: {CurrentSCN=9223372036854775807} Getting iterators for ResultIterators [name=PARALLEL,id=3241642d-a34a-131b-c1ac-bcc1241482c0,scans=[[{"loadColumnFamiliesOnDemand":true,"filter":"LINK_TYPE = 1","startRow":"","stopRow":"","batch":-1,"cacheBlocks":true,"totalColumns":1,"maxResultSize":-1,"families":{"0":["LINK_TYPE"]},"caching":2147483647,"maxVersions":1,"timeRange":[0,9223372036854775807]}]]] 2020-01-30 13:20:57,563 DEBUG org.apache.phoenix.iterate.ParallelIterators: {CurrentSCN=9223372036854775807} Id: 3241642d-a34a-131b-c1ac-bcc1241482c0, Time: 0ms, Scan: {"loadColumnFamiliesOnDemand":true,"filter":"LINK_TYPE = 1","startRow":"","stopRow":"","batch":-1,"cacheBlocks":true,"totalColumns":1,"maxResultSize":-1,"families":{"0":["LINK_TYPE"]},"caching":2147483647,"maxVersions":1,"timeRange":[0,9223372036854775807]} 2020-01-30 13:20:57,600 INFO org.apache.phoenix.util.UpgradeUtil: Upgrading metadata to add parent to child links for views 2020-01-30 13:20:57,604 DEBUG org.apache.phoenix.jdbc.PhoenixStatement: Reloading table CHILD_LINK data from server 2020-01-30 13:20:57,619 DEBUG org.apache.phoenix.query.ConnectionQueryServicesImpl: 8406@hbase-master-upgrade001-stg released mutex for tenantId : null schemaName : SYSTEM tableName : CATALOG columnName : null familyName : null 2020-01-30 13:20:57,619 INFO org.apache.hadoop.hbase.client.ConnectionImplementation: Closing master protocol: MasterService 2020-01-30 13:20:57,623 DEBUG org.apache.phoenix.shaded.org.apache.zookeeper.ZooKeeper: Closing session: 0x16fece700a300fb 2020-01-30 13:20:57,623 DEBUG org.apache.phoenix.shaded.org.apache.zookeeper.ClientCnxn: Closing client for session: 0x16fece700a300fb 2020-01-30 13:20:57,623 DEBUG org.apache.phoenix.jdbc.PhoenixDriver: Expiring hbase-data-upgrade001-stg.foo.bar,hbase-data-upgrade002-stg.foo.bar,hbase-master-upgrade001-stg.foo.bar:2181:/hbase because of EXPLICIT 2020-01-30 13:20:57,623 INFO org.apache.phoenix.log.QueryLoggerDisruptor: Shutting down QueryLoggerDisruptor.. 2020-01-30 13:20:57,627 DEBUG org.apache.phoenix.shaded.org.eclipse.jetty.server.Server: RESPONSE / 500 handled=true As for Phoenix - I used version built by our devs. To make sure that everything is right, I want to use the official version, but I can't find tarbal on https://phoenix.apache.org/download.html page. On Wed, Jan 29, 2020 at 6:55 PM Josh Elser <els...@apache.org> wrote: > Aleksandr and Prathap, > > Upgrades are done in Phoenix as they always have been. You should deploy > the new version of phoenix-server jars to HBase, and then the first time > a client connects with the Phoenix JDBC driver, that client will trigger > an update to any system tables schema. > > As such, you need to make sure that this client has permission to alter > the phoenix system tables that exist, often requiring admin-level access > to hbase. Your first step should be collecting DEBUG log from your > Phoenix JDBC client on upgrade. > > Please also remember that 5.0.0 is pretty old at this point -- we're > overdue for a 5.1.0. There may be existing issues that have already been > fixed around the upgrade. Doing a search on Jira if you've not done so > already is important. > > On 1/29/20 4:30 AM, Aleksandr Saraseka wrote: > > Hello. > > I'm second on this. > > We upgraded phoenix from 4.14.0 to 5.0.0 (with all underlying things > > like hdfs, hbase) and have the same problem. > > > > We are using queryserver + thin-client > > So on PQS side we have: > > 2020-01-29 09:24:21,579 INFO org.apache.phoenix.util.UpgradeUtil: > > Upgrading metadata to add parent links for indexes on views > > 2020-01-29 09:24:21,615 INFO org.apache.phoenix.util.UpgradeUtil: > > Upgrading metadata to add parent to child links for views > > 2020-01-29 09:24:21,628 INFO > > org.apache.hadoop.hbase.client.ConnectionImplementation: Closing master > > protocol: MasterService > > 2020-01-29 09:24:21,631 INFO > > org.apache.phoenix.log.QueryLoggerDisruptor: Shutting down > > QueryLoggerDisruptor.. > > > > On client side: > > java.lang.RuntimeException: > > org.apache.phoenix.schema.TableNotFoundException: ERROR 1012 (42M03): > > Table undefined. tableName=SYSTEM.CHILD_LINK > > > > Can you point me to upgrade guide for Phoenix ? I tried to find it by > > myself and have no luck. > > > > On Thu, Jan 16, 2020 at 1:08 PM Prathap Rajendran <prathap...@gmail.com > > <mailto:prathap...@gmail.com>> wrote: > > > > Hi All, > > > > Thanks for the quick update. Still we have some clarification about > > the context. > > > > Actually we are upgrading from the below version > > Source : apache-phoenix-4.14.0-cdh5.14.2 > > Destination: apache-phoenix-5.0.0-HBase-2.0-bin.tar.gz > > < > http://csfci.ih.lucent.com/~prathapr/phoenix62/apache-phoenix-5.0.0-HBase-2.0-bin.tar.gz > > > > > > Just FYI, we have already upgraded to Hbase 2.0. > > > > Still we are facing the issue below, Once we create this table > > manually, then there is no issues to run DML operations. > > > org.apache.hadoop.hbase.TableNotFoundException: > > SYSTEM.CHILD_LINK > > > > Please let me know if any steps/documents for phoenix upgrade from > > 4.14 to 5.0. > > > > Thanks, > > Prathap > > > > > > On Tue, Jan 14, 2020 at 11:34 PM Josh Elser <els...@apache.org > > <mailto:els...@apache.org>> wrote: > > > > (with VP-Phoenix hat on) > > > > This is not an official Apache Phoenix release, nor does it > > follow the > > ASF trademarks/branding rules. I'll be following up with the > > author to > > address the trademark violations. > > > > Please direct your questions to the author of this project. > > Again, it is > > *not* Apache Phoenix. > > > > On 1/14/20 12:37 PM, Geoffrey Jacoby wrote: > > > Phoenix 5.1 doesn't actually exist yet, at least not at the > > Apache > > > level. We haven't released it yet. It's possible that a > > vendor or user > > > has cut an unofficial release off one of our > > development branches, but > > > that's not something we can give support on. You should > > contact your > > > vendor. > > > > > > Also, since I see you're upgrading from Phoenix 4.14 to 5.1: > > The 4.x > > > branch of Phoenix is for HBase 1.x systems, and the 5.x > > branch is for > > > HBase 2.x systems. If you're upgrading from a 4.x to a 5.x, > > make sure > > > that you also upgrade your HBase. If you're still on HBase > > 1.x, we > > > recently released Phoenix 4.15, which does have a supported > > upgrade path > > > from 4.14 (and a very similar set of features to what 5.1 will > > > eventually get). > > > > > > Geoffrey > > > > > > On Tue, Jan 14, 2020 at 5:23 AM Prathap Rajendran > > <prathap...@gmail.com <mailto:prathap...@gmail.com> > > > <mailto:prathap...@gmail.com <mailto:prathap...@gmail.com>>> > > wrote: > > > > > > Hello All, > > > > > > We are trying to upgrade the phoenix version from > > > "apache-phoenix-4.14.0-cdh5.14.2" to > > "APACHE_PHOENIX-5.1.0-cdh6.1.0." > > > > > > I couldn't find out any upgrade steps for the same. > > Please help me > > > out to get any documents available. > > > *_Note:_* > > > I have downloaded the below phoenix parcel and trying to > > access some > > > DML operation. I am getting the following error > > > > > > > > > https://github.com/dmilan77/cloudera-phoenix/releases/download/5.1.0-HBase-2.0-cdh6.1.0/APACHE_PHOENIX-5.1.0-cdh6.1.0.p1.0-el7.parcel > > < > https://github.com/dmilan77/cloudera-phoenix/releases/download/5.1.0-HBase-2.0-cdh6.1.0/APACHE_PHOENIX-5.1.0-cdh6.1.0.p1.0-el7.parcel > > > > > > > < > https://github.com/dmilan77/cloudera-phoenix/releases/download/5.1.0-HBase-2.0-cdh6.1.0/APACHE_PHOENIX-5.1.0-cdh6.1.0.p1.0-el7.parcel > < > https://github.com/dmilan77/cloudera-phoenix/releases/download/5.1.0-HBase-2.0-cdh6.1.0/APACHE_PHOENIX-5.1.0-cdh6.1.0.p1.0-el7.parcel > >> > > > > > > *_Error:_* > > > 20/01/13 04:22:41 WARN client.HTable: Error calling > > coprocessor > > > service > > > > > > org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService > > > for row \x00\x00WEB_STAT > > > java.util.concurrent.ExecutionException: > > > org.apache.hadoop.hbase.TableNotFoundException: > > > org.apache.hadoop.hbase.TableNotFoundException: > > SYSTEM.CHILD_LINK > > > at > > > > > > > org.apache.hadoop.hbase.client.ConnectionImplementation.locateRegionInMeta(ConnectionImplementation.java:860) > > > at > > > > > > > org.apache.hadoop.hbase.client.ConnectionImplementation.locateRegion(ConnectionImplementation.java:755) > > > at > > > > > > > org.apache.hadoop.hbase.client.ConnectionUtils$ShortCircuitingClusterConnection.locateRegion(ConnectionUtils.java:137) > > > at > > > > > > > org.apache.hadoop.hbase.client.RpcRetryingCallerWithReadReplicas.getRegionLocations(RpcRetryingCallerWithReadReplicas.java:326) > > > at > > > > > > > org.apache.hadoop.hbase.client.ScannerCallableWithReplicas.call(ScannerCallableWithReplicas.java:153) > > > at > > > > > > > org.apache.hadoop.hbase.client.ScannerCallableWithReplicas.call(ScannerCallableWithReplicas.java:58) > > > at > > > > > > > org.apache.hadoop.hbase.client.RpcRetryingCallerImpl.callWithoutRetries(RpcRetryingCallerImpl.java:192) > > > at > > > > > > org.apache.hadoop.hbase.client.ClientScanner.call(ClientScanner.java:267) > > > at > > > > > > > org.apache.hadoop.hbase.client.ClientScanner.loadCache(ClientScanner.java:435) > > > at > > > > > > > org.apache.hadoop.hbase.client.ClientScanner.nextWithSyncCache(ClientScanner.java:310) > > > at > > > > > > org.apache.hadoop.hbase.client.ClientScanner.next(ClientScanner.java:595) > > > at > > > > > > > org.apache.phoenix.coprocessor.ViewFinder.findRelatedViews(ViewFinder.java:94) > > > at > > > > > > > org.apache.phoenix.coprocessor.MetaDataEndpointImpl.dropChildViews(MetaDataEndpointImpl.java:2488) > > > at > > > > > > > org.apache.phoenix.coprocessor.MetaDataEndpointImpl.createTable(MetaDataEndpointImpl.java:2083) > > > at > > > > > > > org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService.callMethod(MetaDataProtos.java:17053) > > > at > > > > > > org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:8218) > > > at > > > > > > > org.apache.hadoop.hbase.regionserver.RSRpcServices.execServiceOnRegion(RSRpcServices.java:2423) > > > at > > > > > > > org.apache.hadoop.hbase.regionserver.RSRpcServices.execService(RSRpcServices.java:2405) > > > at > > > > > > > org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:42010)*_ > > > _* > > > > > > Thanks, > > > Prathap > > > > > > > > > > > -- > > Aleksandr Saraseka > > DBA > > 380997600401 > > <tel:380997600401> *•* asaras...@eztexting.com > > <mailto:asaras...@eztexting.com> *•* eztexting.com > > < > http://eztexting.com?utm_source=WiseStamp&utm_medium=email&utm_term=&utm_content=&utm_campaign=signature> > > > > > > > < > http://facebook.com/eztexting?utm_source=WiseStamp&utm_medium=email&utm_term=&utm_content=&utm_campaign=signature> > > > < > http://linkedin.com/company/eztexting/?utm_source=WiseStamp&utm_medium=email&utm_term=&utm_content=&utm_campaign=signature> > > > < > http://twitter.com/eztexting?utm_source=WiseStamp&utm_medium=email&utm_term=&utm_content=&utm_campaign=signature> > > > < > https://www.youtube.com/eztexting?utm_source=WiseStamp&utm_medium=email&utm_term=&utm_content=&utm_campaign=signature> > > > < > https://www.instagram.com/ez_texting/?utm_source=WiseStamp&utm_medium=email&utm_term=&utm_content=&utm_campaign=signature> > > > < > https://www.facebook.com/alex.saraseka?utm_source=WiseStamp&utm_medium=email&utm_term=&utm_content=&utm_campaign=signature> > > > < > https://www.linkedin.com/in/alexander-saraseka-32616076/?utm_source=WiseStamp&utm_medium=email&utm_term=&utm_content=&utm_campaign=signature > > > > > -- Aleksandr Saraseka DBA 380997600401 *•* asaras...@eztexting.com *•* eztexting.com <http://eztexting.com?utm_source=WiseStamp&utm_medium=email&utm_term=&utm_content=&utm_campaign=signature> <http://facebook.com/eztexting?utm_source=WiseStamp&utm_medium=email&utm_term=&utm_content=&utm_campaign=signature> <http://linkedin.com/company/eztexting/?utm_source=WiseStamp&utm_medium=email&utm_term=&utm_content=&utm_campaign=signature> <http://twitter.com/eztexting?utm_source=WiseStamp&utm_medium=email&utm_term=&utm_content=&utm_campaign=signature> <https://www.youtube.com/eztexting?utm_source=WiseStamp&utm_medium=email&utm_term=&utm_content=&utm_campaign=signature> <https://www.instagram.com/ez_texting/?utm_source=WiseStamp&utm_medium=email&utm_term=&utm_content=&utm_campaign=signature> <https://www.facebook.com/alex.saraseka?utm_source=WiseStamp&utm_medium=email&utm_term=&utm_content=&utm_campaign=signature> <https://www.linkedin.com/in/alexander-saraseka-32616076/?utm_source=WiseStamp&utm_medium=email&utm_term=&utm_content=&utm_campaign=signature>