Hello.
I'm second on this.
We upgraded phoenix from 4.14.0 to 5.0.0 (with all underlying things like
hdfs, hbase) and have the same problem.

We are using queryserver + thin-client
So on PQS side we have:
2020-01-29 09:24:21,579 INFO org.apache.phoenix.util.UpgradeUtil: Upgrading
metadata to add parent links for indexes on views
2020-01-29 09:24:21,615 INFO org.apache.phoenix.util.UpgradeUtil: Upgrading
metadata to add parent to child links for views
2020-01-29 09:24:21,628 INFO
org.apache.hadoop.hbase.client.ConnectionImplementation: Closing master
protocol: MasterService
2020-01-29 09:24:21,631 INFO org.apache.phoenix.log.QueryLoggerDisruptor:
Shutting down QueryLoggerDisruptor..

On client side:
java.lang.RuntimeException:
org.apache.phoenix.schema.TableNotFoundException: ERROR 1012 (42M03): Table
undefined. tableName=SYSTEM.CHILD_LINK

Can you point me to upgrade guide for Phoenix ? I tried to find it by
myself and have no luck.

On Thu, Jan 16, 2020 at 1:08 PM Prathap Rajendran <prathap...@gmail.com>
wrote:

> Hi All,
>
> Thanks for the quick update. Still we have some clarification about the
> context.
>
> Actually we are upgrading from the below version
> Source      : apache-phoenix-4.14.0-cdh5.14.2
> Destination: apache-phoenix-5.0.0-HBase-2.0-bin.tar.gz
> <http://csfci.ih.lucent.com/~prathapr/phoenix62/apache-phoenix-5.0.0-HBase-2.0-bin.tar.gz>
>
> Just FYI, we have already upgraded to Hbase  2.0.
>
> Still we are facing the issue below,  Once we create this table manually,
> then there is no issues to run DML operations.
>   >     org.apache.hadoop.hbase.TableNotFoundException: SYSTEM.CHILD_LINK
>
> Please let me know if any steps/documents for phoenix upgrade from 4.14 to
> 5.0.
>
> Thanks,
> Prathap
>
>
> On Tue, Jan 14, 2020 at 11:34 PM Josh Elser <els...@apache.org> wrote:
>
>> (with VP-Phoenix hat on)
>>
>> This is not an official Apache Phoenix release, nor does it follow the
>> ASF trademarks/branding rules. I'll be following up with the author to
>> address the trademark violations.
>>
>> Please direct your questions to the author of this project. Again, it is
>> *not* Apache Phoenix.
>>
>> On 1/14/20 12:37 PM, Geoffrey Jacoby wrote:
>> > Phoenix 5.1 doesn't actually exist yet, at least not at the Apache
>> > level. We haven't released it yet. It's possible that a vendor or user
>> > has cut an unofficial release off one of our development branches, but
>> > that's not something we can give support on. You should contact your
>> > vendor.
>> >
>> > Also, since I see you're upgrading from Phoenix 4.14 to 5.1: The 4.x
>> > branch of Phoenix is for HBase 1.x systems, and the 5.x branch is for
>> > HBase 2.x systems. If you're upgrading from a 4.x to a 5.x, make sure
>> > that you also upgrade your HBase. If you're still on HBase 1.x, we
>> > recently released Phoenix 4.15, which does have a supported upgrade
>> path
>> > from 4.14 (and a very similar set of features to what 5.1 will
>> > eventually get).
>> >
>> > Geoffrey
>> >
>> > On Tue, Jan 14, 2020 at 5:23 AM Prathap Rajendran <prathap...@gmail.com
>> > <mailto:prathap...@gmail.com>> wrote:
>> >
>> >     Hello All,
>> >
>> >     We are trying to upgrade the phoenix version from
>> >     "apache-phoenix-4.14.0-cdh5.14.2" to
>> "APACHE_PHOENIX-5.1.0-cdh6.1.0."
>> >
>> >     I couldn't find out any upgrade steps for the same. Please help me
>> >     out to get any documents available.
>> >     *_Note:_*
>> >     I have downloaded the below phoenix parcel and trying to access some
>> >     DML operation. I am getting the following error
>> >
>> >
>> https://github.com/dmilan77/cloudera-phoenix/releases/download/5.1.0-HBase-2.0-cdh6.1.0/APACHE_PHOENIX-5.1.0-cdh6.1.0.p1.0-el7.parcel
>> >     <
>> https://github.com/dmilan77/cloudera-phoenix/releases/download/5.1.0-HBase-2.0-cdh6.1.0/APACHE_PHOENIX-5.1.0-cdh6.1.0.p1.0-el7.parcel
>> >
>> >
>> >     *_Error:_*
>> >     20/01/13 04:22:41 WARN client.HTable: Error calling coprocessor
>> >     service
>> >
>>  org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService
>> >     for row \x00\x00WEB_STAT
>> >     java.util.concurrent.ExecutionException:
>> >     org.apache.hadoop.hbase.TableNotFoundException:
>> >     org.apache.hadoop.hbase.TableNotFoundException: SYSTEM.CHILD_LINK
>> >              at
>> >
>>  
>> org.apache.hadoop.hbase.client.ConnectionImplementation.locateRegionInMeta(ConnectionImplementation.java:860)
>> >              at
>> >
>>  
>> org.apache.hadoop.hbase.client.ConnectionImplementation.locateRegion(ConnectionImplementation.java:755)
>> >              at
>> >
>>  
>> org.apache.hadoop.hbase.client.ConnectionUtils$ShortCircuitingClusterConnection.locateRegion(ConnectionUtils.java:137)
>> >              at
>> >
>>  
>> org.apache.hadoop.hbase.client.RpcRetryingCallerWithReadReplicas.getRegionLocations(RpcRetryingCallerWithReadReplicas.java:326)
>> >              at
>> >
>>  
>> org.apache.hadoop.hbase.client.ScannerCallableWithReplicas.call(ScannerCallableWithReplicas.java:153)
>> >              at
>> >
>>  
>> org.apache.hadoop.hbase.client.ScannerCallableWithReplicas.call(ScannerCallableWithReplicas.java:58)
>> >              at
>> >
>>  
>> org.apache.hadoop.hbase.client.RpcRetryingCallerImpl.callWithoutRetries(RpcRetryingCallerImpl.java:192)
>> >              at
>> >
>>  org.apache.hadoop.hbase.client.ClientScanner.call(ClientScanner.java:267)
>> >              at
>> >
>>  
>> org.apache.hadoop.hbase.client.ClientScanner.loadCache(ClientScanner.java:435)
>> >              at
>> >
>>  
>> org.apache.hadoop.hbase.client.ClientScanner.nextWithSyncCache(ClientScanner.java:310)
>> >              at
>> >
>>  org.apache.hadoop.hbase.client.ClientScanner.next(ClientScanner.java:595)
>> >              at
>> >
>>  
>> org.apache.phoenix.coprocessor.ViewFinder.findRelatedViews(ViewFinder.java:94)
>> >              at
>> >
>>  
>> org.apache.phoenix.coprocessor.MetaDataEndpointImpl.dropChildViews(MetaDataEndpointImpl.java:2488)
>> >              at
>> >
>>  
>> org.apache.phoenix.coprocessor.MetaDataEndpointImpl.createTable(MetaDataEndpointImpl.java:2083)
>> >              at
>> >
>>  
>> org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService.callMethod(MetaDataProtos.java:17053)
>> >              at
>> >
>>  org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:8218)
>> >              at
>> >
>>  
>> org.apache.hadoop.hbase.regionserver.RSRpcServices.execServiceOnRegion(RSRpcServices.java:2423)
>> >              at
>> >
>>  
>> org.apache.hadoop.hbase.regionserver.RSRpcServices.execService(RSRpcServices.java:2405)
>> >              at
>> >
>>  
>> org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:42010)*_
>> >     _*
>> >
>> >     Thanks,
>> >     Prathap
>> >
>>
>

-- 
Aleksandr Saraseka
DBA
380997600401
 *•*  asaras...@eztexting.com  *•*  eztexting.com
<http://eztexting.com?utm_source=WiseStamp&utm_medium=email&utm_term=&utm_content=&utm_campaign=signature>
<http://facebook.com/eztexting?utm_source=WiseStamp&utm_medium=email&utm_term=&utm_content=&utm_campaign=signature>
<http://linkedin.com/company/eztexting/?utm_source=WiseStamp&utm_medium=email&utm_term=&utm_content=&utm_campaign=signature>
<http://twitter.com/eztexting?utm_source=WiseStamp&utm_medium=email&utm_term=&utm_content=&utm_campaign=signature>
<https://www.youtube.com/eztexting?utm_source=WiseStamp&utm_medium=email&utm_term=&utm_content=&utm_campaign=signature>
<https://www.instagram.com/ez_texting/?utm_source=WiseStamp&utm_medium=email&utm_term=&utm_content=&utm_campaign=signature>
<https://www.facebook.com/alex.saraseka?utm_source=WiseStamp&utm_medium=email&utm_term=&utm_content=&utm_campaign=signature>
<https://www.linkedin.com/in/alexander-saraseka-32616076/?utm_source=WiseStamp&utm_medium=email&utm_term=&utm_content=&utm_campaign=signature>

Reply via email to