Hi All,

Thanks for the quick update. Still we have some clarification about the
context.

Actually we are upgrading from the below version
Source      : apache-phoenix-4.14.0-cdh5.14.2
Destination: apache-phoenix-5.0.0-HBase-2.0-bin.tar.gz
<http://csfci.ih.lucent.com/~prathapr/phoenix62/apache-phoenix-5.0.0-HBase-2.0-bin.tar.gz>

Just FYI, we have already upgraded to Hbase  2.0.

Still we are facing the issue below,  Once we create this table manually,
then there is no issues to run DML operations.
  >     org.apache.hadoop.hbase.TableNotFoundException: SYSTEM.CHILD_LINK

Please let me know if any steps/documents for phoenix upgrade from 4.14 to
5.0.

Thanks,
Prathap


On Tue, Jan 14, 2020 at 11:34 PM Josh Elser <els...@apache.org> wrote:

> (with VP-Phoenix hat on)
>
> This is not an official Apache Phoenix release, nor does it follow the
> ASF trademarks/branding rules. I'll be following up with the author to
> address the trademark violations.
>
> Please direct your questions to the author of this project. Again, it is
> *not* Apache Phoenix.
>
> On 1/14/20 12:37 PM, Geoffrey Jacoby wrote:
> > Phoenix 5.1 doesn't actually exist yet, at least not at the Apache
> > level. We haven't released it yet. It's possible that a vendor or user
> > has cut an unofficial release off one of our development branches, but
> > that's not something we can give support on. You should contact your
> > vendor.
> >
> > Also, since I see you're upgrading from Phoenix 4.14 to 5.1: The 4.x
> > branch of Phoenix is for HBase 1.x systems, and the 5.x branch is for
> > HBase 2.x systems. If you're upgrading from a 4.x to a 5.x, make sure
> > that you also upgrade your HBase. If you're still on HBase 1.x, we
> > recently released Phoenix 4.15, which does have a supported upgrade path
> > from 4.14 (and a very similar set of features to what 5.1 will
> > eventually get).
> >
> > Geoffrey
> >
> > On Tue, Jan 14, 2020 at 5:23 AM Prathap Rajendran <prathap...@gmail.com
> > <mailto:prathap...@gmail.com>> wrote:
> >
> >     Hello All,
> >
> >     We are trying to upgrade the phoenix version from
> >     "apache-phoenix-4.14.0-cdh5.14.2" to "APACHE_PHOENIX-5.1.0-cdh6.1.0."
> >
> >     I couldn't find out any upgrade steps for the same. Please help me
> >     out to get any documents available.
> >     *_Note:_*
> >     I have downloaded the below phoenix parcel and trying to access some
> >     DML operation. I am getting the following error
> >
> >
> https://github.com/dmilan77/cloudera-phoenix/releases/download/5.1.0-HBase-2.0-cdh6.1.0/APACHE_PHOENIX-5.1.0-cdh6.1.0.p1.0-el7.parcel
> >     <
> https://github.com/dmilan77/cloudera-phoenix/releases/download/5.1.0-HBase-2.0-cdh6.1.0/APACHE_PHOENIX-5.1.0-cdh6.1.0.p1.0-el7.parcel
> >
> >
> >     *_Error:_*
> >     20/01/13 04:22:41 WARN client.HTable: Error calling coprocessor
> >     service
> >
>  org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService
> >     for row \x00\x00WEB_STAT
> >     java.util.concurrent.ExecutionException:
> >     org.apache.hadoop.hbase.TableNotFoundException:
> >     org.apache.hadoop.hbase.TableNotFoundException: SYSTEM.CHILD_LINK
> >              at
> >
>  
> org.apache.hadoop.hbase.client.ConnectionImplementation.locateRegionInMeta(ConnectionImplementation.java:860)
> >              at
> >
>  
> org.apache.hadoop.hbase.client.ConnectionImplementation.locateRegion(ConnectionImplementation.java:755)
> >              at
> >
>  
> org.apache.hadoop.hbase.client.ConnectionUtils$ShortCircuitingClusterConnection.locateRegion(ConnectionUtils.java:137)
> >              at
> >
>  
> org.apache.hadoop.hbase.client.RpcRetryingCallerWithReadReplicas.getRegionLocations(RpcRetryingCallerWithReadReplicas.java:326)
> >              at
> >
>  
> org.apache.hadoop.hbase.client.ScannerCallableWithReplicas.call(ScannerCallableWithReplicas.java:153)
> >              at
> >
>  
> org.apache.hadoop.hbase.client.ScannerCallableWithReplicas.call(ScannerCallableWithReplicas.java:58)
> >              at
> >
>  
> org.apache.hadoop.hbase.client.RpcRetryingCallerImpl.callWithoutRetries(RpcRetryingCallerImpl.java:192)
> >              at
> >
>  org.apache.hadoop.hbase.client.ClientScanner.call(ClientScanner.java:267)
> >              at
> >
>  
> org.apache.hadoop.hbase.client.ClientScanner.loadCache(ClientScanner.java:435)
> >              at
> >
>  
> org.apache.hadoop.hbase.client.ClientScanner.nextWithSyncCache(ClientScanner.java:310)
> >              at
> >
>  org.apache.hadoop.hbase.client.ClientScanner.next(ClientScanner.java:595)
> >              at
> >
>  
> org.apache.phoenix.coprocessor.ViewFinder.findRelatedViews(ViewFinder.java:94)
> >              at
> >
>  
> org.apache.phoenix.coprocessor.MetaDataEndpointImpl.dropChildViews(MetaDataEndpointImpl.java:2488)
> >              at
> >
>  
> org.apache.phoenix.coprocessor.MetaDataEndpointImpl.createTable(MetaDataEndpointImpl.java:2083)
> >              at
> >
>  
> org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService.callMethod(MetaDataProtos.java:17053)
> >              at
> >
>  org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:8218)
> >              at
> >
>  
> org.apache.hadoop.hbase.regionserver.RSRpcServices.execServiceOnRegion(RSRpcServices.java:2423)
> >              at
> >
>  
> org.apache.hadoop.hbase.regionserver.RSRpcServices.execService(RSRpcServices.java:2405)
> >              at
> >
>  
> org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:42010)*_
> >     _*
> >
> >     Thanks,
> >     Prathap
> >
>

Reply via email to