"Phoenix-server" refers to the phoenix-$VERSION-server.jar that is either included in the binary tarball or is generated by the official source-release.

"Deploying" it means copying the jar to $HBASE_HOME/lib.

On 8/6/18 9:56 PM, 倪项菲 wrote:

Hi Zhang Yun,
    the link you mentioned tells us to add the phoenix jar to  hbase lib directory,it doesn't tell us how to deploy the phoenix server.

    发件人: Jaanai Zhang <mailto:cloud.pos...@gmail.com>
    时间: 2018/08/07(星期二)09:36
    收件人: user <mailto:user@phoenix.apache.org>;
    主题: Re: Re: error when using apache-phoenix-4.14.0-HBase-1.2-bin
    with hbase 1.2.6

reference link: http://phoenix.apache.org/installation.html


----------------------------------------
    Yun Zhang
    Best regards!


2018-08-07 9:30 GMT+08:00 倪项菲 <nixiangfei_...@chinamobile.com <mailto:nixiangfei_...@chinamobile.com>>:

    Hi Zhang Yun,
         how to deploy the Phoenix server?I just have the infomation
    from phoenix website,it doesn't mention the phoenix server

        发件人: Jaanai Zhang <mailto:cloud.pos...@gmail.com>
        时间: 2018/08/07(星期二)09:16
        收件人: user <mailto:user@phoenix.apache.org>;
        主题: Re: error when using apache-phoenix-4.14.0-HBase-1.2-bin
        with hbase 1.2.6

    Please ensure your Phoenix server was deployed and had resarted


    ----------------------------------------
        Yun Zhang
        Best regards!


    2018-08-07 9:10 GMT+08:00 倪项菲 <nixiangfei_...@chinamobile.com
    <mailto:nixiangfei_...@chinamobile.com>>:


        Hi Experts,
             I am using HBase 1.2.6,the cluster is working good with
        HMaster HA,but when we integrate phoenix with hbase,it
        failed,below are the steps
             1,download apache-phoenix-4.14.0-HBase-1.2-bin from
        http://phoenix.apache.org,the copy the tar file to the HMaster
        and unzip the file
2,copy phoenix-core-4.14.0-HBase-1.2.jar phoenix-4.14.0-HBase-1.2-server.jar
        to all HBase nodes including HMaster and HRegionServer ,put them
        to hbasehome/lib,my path is /opt/hbase-1.2.6/lib
             3,restart hbase cluster
             4,then start to use phoenix,but it return below error:
        [apache@plat-ecloud01-bigdata-journalnode01 bin]$ ./sqlline.py
        
plat-ecloud01-bigdata-zk01,plat-ecloud01-bigdata-zk02,plat-ecloud01-bigdata-zk03
        Setting property: [incremental, false]
        Setting property: [isolation, TRANSACTION_READ_COMMITTED]
        issuing: !connect jdbc:phoenix:plat-ecloud01-bigdata-zk01 none
        none org.apache.phoenix.jdbc.PhoenixDriver
        Connecting to
        
jdbc:phoenix:plat-ecloud01-bigdata-zk01,plat-ecloud01-bigdata-zk02,plat-ecloud01-bigdata-zk03
        SLF4J: Class path contains multiple SLF4J bindings.
        SLF4J: Found binding in
        
[jar:file:/opt/apache-phoenix-4.14.0-HBase-1.2-bin/phoenix-4.14.0-HBase-1.2-client.jar!/org/slf4j/impl/StaticLoggerBinder.class]
        SLF4J: Found binding in
        
[jar:file:/opt/hadoop-2.7.6/share/hadoop/common/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class]
        SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings
        <http://www.slf4j.org/codes.html#multiple_bindings> for an
        explanation.
        18/08/06 18:40:08 WARN util.NativeCodeLoader: Unable to load
        native-hadoop library for your platform... using builtin-java
        classes where applicable
        Error: org.apache.hadoop.hbase.DoNotRetryIOException: Unable to
        load configured region split policy
        'org.apache.phoenix.schema.MetaDataSplitPolicy' for table
        'SYSTEM.CATALOG' Set hbase.table.sanity.checks to false at conf
        or table descriptor if you want to bypass sanity checks
                 at
        
org.apache.hadoop.hbase.master.HMaster.warnOrThrowExceptionForFailure(HMaster.java:1754)
                 at
        
org.apache.hadoop.hbase.master.HMaster.sanityCheckTableDescriptor(HMaster.java:1615)
                 at
        org.apache.hadoop.hbase.master.HMaster.createTable(HMaster.java:1541)
                 at
        
org.apache.hadoop.hbase.master.MasterRpcServices.createTable(MasterRpcServices.java:463)
                 at
        
org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:55682)
                 at
        org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2196)
                 at org.apache.hadoop.hbase.ipc.Ca
        <http://org.apache.hadoop.hbase.ipc.Ca>llRunner.run(CallRunner.java:112)
                 at
        
org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)
                 at
        org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)
                 at java.lang.Thread.run(Thread.java:745)
        (state=08000,code=101)
        org.apache.phoenix.exception.PhoenixIOException:
        org.apache.hadoop.hbase.DoNotRetryIOException: Unable to load
        configured region split policy
        'org.apache.phoenix.schema.MetaDataSplitPolicy' for table
        'SYSTEM.CATALOG' Set hbase.table.sanity.checks to false at conf
        or table descriptor if you want to bypass sanity checks
                 at
        
org.apache.hadoop.hbase.master.HMaster.warnOrThrowExceptionForFailure(HMaster.java:1754)
                 at
        
org.apache.hadoop.hbase.master.HMaster.sanityCheckTableDescriptor(HMaster.java:1615)
                 at
        org.apache.hadoop.hbase.master.HMaster.createTable(HMaster.java:1541)
                 at
        
org.apache.hadoop.hbase.master.MasterRpcServices.createTable(MasterRpcServices.java:463)
                 at
        
org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:55682)
                 at
        org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2196)
                 at org.apache.hadoop.hbase.ipc.Ca
        <http://org.apache.hadoop.hbase.ipc.Ca>llRunner.run(CallRunner.java:112)
                 at
        
org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)
                 at
        org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)
                 at java.lang.Thread.run(Thread.java:745)

                 at
        
org.apache.phoenix.util.ServerUtil.parseServerException(ServerUtil.java:144)
                 at
        
org.apache.phoenix.query.ConnectionQueryServicesImpl.ensureTableCreated(ConnectionQueryServicesImpl.java:1197)
                 at
        
org.apache.phoenix.query.ConnectionQueryServicesImpl.createTable(ConnectionQueryServicesImpl.java:1491)
                 at
        
org.apache.phoenix.schema.MetaDataClient.createTableInternal(MetaDataClient.java:2717)
                 at
        
org.apache.phoenix.schema.MetaDataClient.createTable(MetaDataClient.java:1114)
                 at
        
org.apache.phoenix.compile.CreateTableCompiler$1.execute(CreateTableCompiler.java:192)
                 at
        
org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:408)
                 at
        
org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:391)
                 at
        org.apache.phoenix.call.CallRunner.run(CallRunner.java:53)
                 at
        
org.apache.phoenix.jdbc.PhoenixStatement.executeMutation(PhoenixStatement.java:389)
                 at
        
org.apache.phoenix.jdbc.PhoenixStatement.executeMutation(PhoenixStatement.java:378)
                 at
        
org.apache.phoenix.jdbc.PhoenixStatement.executeUpdate(PhoenixStatement.java:1806)
                 at
        
org.apache.phoenix.query.ConnectionQueryServicesImpl$12.call(ConnectionQueryServicesImpl.java:2528)
                 at
        
org.apache.phoenix.query.ConnectionQueryServicesImpl$12.call(ConnectionQueryServicesImpl.java:2491)
                 at
        
org.apache.phoenix.util.PhoenixContextExecutor.call(PhoenixContextExecutor.java:76)
                 at
        
org.apache.phoenix.query.ConnectionQueryServicesImpl.init(ConnectionQueryServicesImpl.java:2491)
                 at
        
org.apache.phoenix.jdbc.PhoenixDriver.getConnectionQueryServices(PhoenixDriver.java:255)
                 at
        
org.apache.phoenix.jdbc.PhoenixEmbeddedDriver.createConnection(PhoenixEmbeddedDriver.java:150)
                 at
        org.apache.phoenix.jdbc.PhoenixDriver.connect(PhoenixDriver.java:221)
                 at
        sqlline.DatabaseConnection.connect(DatabaseConnection.java:157)
                 at
        sqlline.DatabaseConnection.getConnection(DatabaseConnection.java:203)
                 at sqlline.Commands.connect(Commands.java:1064)
                 at sqlline.Commands.connect(Commands.java:996)
                 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native
        Method)
                 at
        
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
                 at
        
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
                 at java.lang.reflect.Method.invoke(Method.java:498)
                 at
        
sqlline.ReflectiveCommandHandler.execute(ReflectiveCommandHandler.java:38)
                 at sqlline.SqlLine.dispatch(SqlLine.java:809)
                 at sqlline.SqlLine.initArgs(SqlLine.java:588)
                 at sqlline.SqlLine.begin(SqlLine.java:661)
                 at sqlline.SqlLine.start(SqlLine.java:398)
                 at sqlline.SqlLine.main(SqlLine.java:291)
        Caused by: org.apache.hadoop.hbase.DoNotRetryIOException:
        org.apache.hadoop.hbase.DoNotRetryIOException: Unable to load
        configured region split policy
        'org.apache.phoenix.schema.MetaDataSplitPolicy' for table
        'SYSTEM.CATALOG' Set hbase.table.sanity.checks to false at conf
        or table descriptor if you want to bypass sanity checks

               I searched from internet,but got no help.
               Any help will be highly appreciated!



Reply via email to