Hi Ankit,




   I have put phoenix-4.14.0-HBase-1.2-server.jar to /opt/hbase-1.2.6/lib,is 
this correct?











 



发件人: Ankit Singhal

时间: 2018/08/14(星期二)02:25

收件人: user;

主题: Re: Re: error when using apache-phoenix-4.14.0-HBase-1.2-bin with hbase 
1.2.6

Skipping sanity checks may unstabilize the functionality on which Phoenix 
relies on, SplitPolicy should have been loaded to prevent splitting of 
SYSTEM.CATALOG table, so to actually fix the issue please check if you have 
right phoenix-server.jar in HBase classpath

"Unable to load configured region split policy 
'org.apache.phoenix.schema.MetaDataSplitPolicy' for table 'SYSTEM.CATALOG' Set 
hbase.table.sanity.checks to false at conf or table descriptor if you want to 
bypass sanity checks"





Regards,

Ankit Singhal




On Sun, Aug 12, 2018 at 6:46 PM, 倪项菲 <nixiangfei_...@chinamobile.com> wrote:




Thanks all.




at last I set hbase.table.sanity.checks to false in hbase-site.xml and restart 
hbase cluster,it works.











 



发件人: Josh Elser

时间: 2018/08/07(星期二)20:58

收件人: user;





主题: Re: error when using apache-phoenix-4.14.0-HBase-1.2-bin with hbase 1.2.6



"Phoenix-server" refers to the phoenix-$VERSION-server.jar that is 
either included in the binary tarball or is generated by the official 
source-release.

"Deploying" it means copying the jar to $HBASE_HOME/lib.

On 8/6/18 9:56 PM, 倪项菲 wrote:
> 
> Hi Zhang Yun,
>      the link you mentioned tells us to add the phoenix jar to  hbase 
> lib directory,it doesn't tell us how to deploy the phoenix server.
> 
>     发件人: Jaanai Zhang <mailto:cloud.pos...@gmail.com>
>     时间: 2018/08/07(星期二)09:36
>     收件人: user <mailto:user@phoenix.apache.org>;
>     主题: Re: Re: error when using apache-phoenix-4.14.0-HBase-1.2-bin
>     with hbase 1.2.6
> 
> reference link: http://phoenix.apache.org/installation.html
> 
> 
> ----------------------------------------
>     Yun Zhang
>     Best regards!
> 
> 
> 2018-08-07 9:30 GMT+08:00 倪项菲 <nixiangfei_...@chinamobile.com 
> <mailto:nixiangfei_...@chinamobile.com>>:
> 
>     Hi Zhang Yun,
>          how to deploy the Phoenix server?I just have the infomation
>     from phoenix website,it doesn't mention the phoenix server
> 
>         发件人: Jaanai Zhang <mailto:cloud.pos...@gmail.com>
>         时间: 2018/08/07(星期二)09:16
>         收件人: user <mailto:user@phoenix.apache.org>;
>         主题: Re: error when using apache-phoenix-4.14.0-HBase-1.2-bin
>         with hbase 1.2.6
> 
>     Please ensure your Phoenix server was deployed and had resarted
> 
> 
>     ----------------------------------------
>         Yun Zhang
>         Best regards!
> 
> 
>     2018-08-07 9:10 GMT+08:00 倪项菲 <nixiangfei_...@chinamobile.com
>     <mailto:nixiangfei_...@chinamobile.com>>:
> 
> 
>         Hi Experts,
>              I am using HBase 1.2.6,the cluster is working good with
>         HMaster HA,but when we integrate phoenix with hbase,it
>         failed,below are the steps
>              1,download apache-phoenix-4.14.0-HBase-1.2-bin from
>         http://phoenix.apache.org,the copy the tar file to the HMaster
>         and unzip the file
>             
>         2,copy phoenix-core-4.14.0-HBase-1.2.jar 
> phoenix-4.14.0-HBase-1.2-server.jar
>         to all HBase nodes including HMaster and HRegionServer ,put them
>         to hbasehome/lib,my path is /opt/hbase-1.2.6/lib
>              3,restart hbase cluster
>              4,then start to use phoenix,but it return below error:
>         [apache@plat-ecloud01-bigdata-journalnode01 bin]$ ./sqlline.py
>         
> plat-ecloud01-bigdata-zk01,plat-ecloud01-bigdata-zk02,plat-ecloud01-bigdata-zk03
>         Setting property: [incremental, false]
>         Setting property: [isolation, TRANSACTION_READ_COMMITTED]
>         issuing: !connect jdbc:phoenix:plat-ecloud01-bigdata-zk01 none
>         none org.apache.phoenix.jdbc.PhoenixDriver
>         Connecting to
>         
> jdbc:phoenix:plat-ecloud01-bigdata-zk01,plat-ecloud01-bigdata-zk02,plat-ecloud01-bigdata-zk03
>         SLF4J: Class path contains multiple SLF4J bindings.
>         SLF4J: Found binding in
>         
> [jar:file:/opt/apache-phoenix-4.14.0-HBase-1.2-bin/phoenix-4.14.0-HBase-1.2-client.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>         SLF4J: Found binding in
>         
> [jar:file:/opt/hadoop-2.7.6/share/hadoop/common/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>         SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings
>         <http://www.slf4j.org/codes.html#multiple_bindings> for an
>         explanation.
>         18/08/06 18:40:08 WARN util.NativeCodeLoader: Unable to load
>         native-hadoop library for your platform... using builtin-java
>         classes where applicable
>         Error: org.apache.hadoop.hbase.DoNotRetryIOException: Unable to
>         load configured region split policy
>         'org.apache.phoenix.schema.MetaDataSplitPolicy' for table
>         'SYSTEM.CATALOG' Set hbase.table.sanity.checks to false at conf
>         or table descriptor if you want to bypass sanity checks
>                  at
>         
> org.apache.hadoop.hbase.master.HMaster.warnOrThrowExceptionForFailure(HMaster.java:1754)
>                  at
>         
> org.apache.hadoop.hbase.master.HMaster.sanityCheckTableDescriptor(HMaster.java:1615)
>                  at
>         org.apache.hadoop.hbase.master.HMaster.createTable(HMaster.java:1541)
>                  at
>         
> org.apache.hadoop.hbase.master.MasterRpcServices.createTable(MasterRpcServices.java:463)
>                  at
>         
> org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:55682)
>                  at
>         org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2196)
>                  at org.apache.hadoop.hbase.ipc.Ca
>         
> <http://org.apache.hadoop.hbase.ipc.Ca>llRunner.run(CallRunner.java:112)
>                  at
>         
> org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)
>                  at
>         org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)
>                  at java.lang.Thread.run(Thread.java:745)
>         (state=08000,code=101)
>         org.apache.phoenix.exception.PhoenixIOException:
>         org.apache.hadoop.hbase.DoNotRetryIOException: Unable to load
>         configured region split policy
>         'org.apache.phoenix.schema.MetaDataSplitPolicy' for table
>         'SYSTEM.CATALOG' Set hbase.table.sanity.checks to false at conf
>         or table descriptor if you want to bypass sanity checks
>                  at
>         
> org.apache.hadoop.hbase.master.HMaster.warnOrThrowExceptionForFailure(HMaster.java:1754)
>                  at
>         
> org.apache.hadoop.hbase.master.HMaster.sanityCheckTableDescriptor(HMaster.java:1615)
>                  at
>         org.apache.hadoop.hbase.master.HMaster.createTable(HMaster.java:1541)
>                  at
>         
> org.apache.hadoop.hbase.master.MasterRpcServices.createTable(MasterRpcServices.java:463)
>                  at
>         
> org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:55682)
>                  at
>         org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2196)
>                  at org.apache.hadoop.hbase.ipc.Ca
>         
> <http://org.apache.hadoop.hbase.ipc.Ca>llRunner.run(CallRunner.java:112)
>                  at
>         
> org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)
>                  at
>         org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)
>                  at java.lang.Thread.run(Thread.java:745)
> 
>                  at
>         
> org.apache.phoenix.util.ServerUtil.parseServerException(ServerUtil.java:144)
>                  at
>         
> org.apache.phoenix.query.ConnectionQueryServicesImpl.ensureTableCreated(ConnectionQueryServicesImpl.java:1197)
>                  at
>         
> org.apache.phoenix.query.ConnectionQueryServicesImpl.createTable(ConnectionQueryServicesImpl.java:1491)
>                  at
>         
> org.apache.phoenix.schema.MetaDataClient.createTableInternal(MetaDataClient.java:2717)
>                  at
>         
> org.apache.phoenix.schema.MetaDataClient.createTable(MetaDataClient.java:1114)
>                  at
>         
> org.apache.phoenix.compile.CreateTableCompiler$1.execute(CreateTableCompiler.java:192)
>                  at
>         
> org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:408)
>                  at
>         
> org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:391)
>                  at
>         org.apache.phoenix.call.CallRunner.run(CallRunner.java:53)
>                  at
>         
> org.apache.phoenix.jdbc.PhoenixStatement.executeMutation(PhoenixStatement.java:389)
>                  at
>         
> org.apache.phoenix.jdbc.PhoenixStatement.executeMutation(PhoenixStatement.java:378)
>                  at
>         
> org.apache.phoenix.jdbc.PhoenixStatement.executeUpdate(PhoenixStatement.java:1806)
>                  at
>         
> org.apache.phoenix.query.ConnectionQueryServicesImpl$12.call(ConnectionQueryServicesImpl.java:2528)
>                  at
>         
> org.apache.phoenix.query.ConnectionQueryServicesImpl$12.call(ConnectionQueryServicesImpl.java:2491)
>                  at
>         
> org.apache.phoenix.util.PhoenixContextExecutor.call(PhoenixContextExecutor.java:76)
>                  at
>         
> org.apache.phoenix.query.ConnectionQueryServicesImpl.init(ConnectionQueryServicesImpl.java:2491)
>                  at
>         
> org.apache.phoenix.jdbc.PhoenixDriver.getConnectionQueryServices(PhoenixDriver.java:255)
>                  at
>         
> org.apache.phoenix.jdbc.PhoenixEmbeddedDriver.createConnection(PhoenixEmbeddedDriver.java:150)
>                  at
>         org.apache.phoenix.jdbc.PhoenixDriver.connect(PhoenixDriver.java:221)
>                  at
>         sqlline.DatabaseConnection.connect(DatabaseConnection.java:157)
>                  at
>         sqlline.DatabaseConnection.getConnection(DatabaseConnection.java:203)
>                  at sqlline.Commands.connect(Commands.java:1064)
>                  at sqlline.Commands.connect(Commands.java:996)
>                  at sun.reflect.NativeMethodAccessorImpl.invoke0(Native
>         Method)
>                  at
>         
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>                  at
>         
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>                  at java.lang.reflect.Method.invoke(Method.java:498)
>                  at
>         
> sqlline.ReflectiveCommandHandler.execute(ReflectiveCommandHandler.java:38)
>                  at sqlline.SqlLine.dispatch(SqlLine.java:809)
>                  at sqlline.SqlLine.initArgs(SqlLine.java:588)
>                  at sqlline.SqlLine.begin(SqlLine.java:661)
>                  at sqlline.SqlLine.start(SqlLine.java:398)
>                  at sqlline.SqlLine.main(SqlLine.java:291)
>         Caused by: org.apache.hadoop.hbase.DoNotRetryIOException:
>         org.apache.hadoop.hbase.DoNotRetryIOException: Unable to load
>         configured region split policy
>         'org.apache.phoenix.schema.MetaDataSplitPolicy' for table
>         'SYSTEM.CATALOG' Set hbase.table.sanity.checks to false at conf
>         or table descriptor if you want to bypass sanity checks
> 
>                I searched from internet,but got no help.
>                Any help will be highly appreciated!
> 
> 
> 
 

Reply via email to