Hi Vikas, Glad you got it working. Just curious - why did you install Phoenix via yum when the HDP 2.1 already comes pre-installed with Phoenix? Thanks, James
On Mon, Sep 1, 2014 at 10:16 AM, Vikas Agarwal <vi...@infoobjects.com> wrote: > Yes, I am using HDP 2.1 and installed Phoenix via yum and it installed 4.0.0 > of phoenix. Added symlink to the phoenix-core*jar into /usr/lib/hbase/lib > and I did restart after these changes. > > However, now, I am able to connect with HBase. Somehow, SYSTEM.CATALOG was > got created in one of my earlier attempts to connect phoenix to HBase and it > didn't contained the required region server information for phoenix to work > on. So, I deleted this table and hope phoenix to recreate it and it did and > connection also worked after that. :) > > > On Mon, Sep 1, 2014 at 6:54 PM, Nicolas Maillard <nmaill...@hortonworks.com> > wrote: >> >> Hello >> Just to be sure you are using HDP 2.1 and you gotten the phoenix jars from >> the page you listed and have put the phoenix jar in all on the hbase nodes >> in the lib directory and restarted the whole hbase service. >> If so could you also paste the line you use to start sqlline. >> >> >> On Mon, Sep 1, 2014 at 12:43 PM, Vikas Agarwal <vi...@infoobjects.com> >> wrote: >>> >>> Hi, >>> >>> We have installed Hadoop cluster using Hortonworks distribution and >>> trying to connect the Phoenix with HBase. However, even after following the >>> steps mentioned here, we are not able to connect Phoenix to HBase. >>> >>> When I am trying to test the connection using sqlline.py, the command >>> hangs the control and I am even not able to do Ctrl-C or Ctrl-Z. After >>> sometime (around 10 min) it throws same exception as with psql.py, >>> complaining mismatch of phoenix jars. >>> >>> When I am trying to test the connection using Phoenix's psql.py command, >>> following set of exceptions are coming: >>> >>> 14/09/01 05:41:25 WARN util.NativeCodeLoader: Unable to load >>> native-hadoop library for your platform... using builtin-java classes where >>> applicable >>> 14/09/01 05:41:26 WARN util.DynamicClassLoader: Failed to identify the fs >>> of dir hdfs://hdp.ambari:8020/apps/hbase/data/lib, ignored >>> java.io.IOException: No FileSystem for scheme: hdfs >>> at >>> org.apache.hadoop.fs.FileSystem.getFileSystemClass(FileSystem.java:2385) >>> at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2392) >>> at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:89) >>> at >>> org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2431) >>> at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2413) >>> at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:368) >>> at org.apache.hadoop.fs.Path.getFileSystem(Path.java:296) >>> at >>> org.apache.hadoop.hbase.util.DynamicClassLoader.<init>(DynamicClassLoader.java:104) >>> at >>> org.apache.hadoop.hbase.protobuf.ProtobufUtil.<clinit>(ProtobufUtil.java:217) >>> at org.apache.hadoop.hbase.ClusterId.parseFrom(ClusterId.java:64) >>> at >>> org.apache.hadoop.hbase.zookeeper.ZKClusterId.readClusterIdZNode(ZKClusterId.java:75) >>> at >>> org.apache.hadoop.hbase.client.ZooKeeperRegistry.getClusterId(ZooKeeperRegistry.java:86) >>> at >>> org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.retrieveClusterId(ConnectionManager.java:853) >>> at >>> org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.<init>(ConnectionManager.java:657) >>> at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) >>> at >>> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57) >>> at >>> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) >>> at java.lang.reflect.Constructor.newInstance(Constructor.java:526) >>> at >>> org.apache.hadoop.hbase.client.ConnectionManager.createConnection(ConnectionManager.java:419) >>> at >>> org.apache.hadoop.hbase.client.ConnectionManager.createConnectionInternal(ConnectionManager.java:314) >>> at >>> org.apache.hadoop.hbase.client.HConnectionManager.createConnection(HConnectionManager.java:291) >>> at >>> org.apache.phoenix.query.HConnectionFactory$HConnectionFactoryImpl.createConnection(HConnectionFactory.java:47) >>> at >>> org.apache.phoenix.query.ConnectionQueryServicesImpl.openConnection(ConnectionQueryServicesImpl.java:252) >>> at >>> org.apache.phoenix.query.ConnectionQueryServicesImpl.init(ConnectionQueryServicesImpl.java:1447) >>> at >>> org.apache.phoenix.jdbc.PhoenixDriver.getConnectionQueryServices(PhoenixDriver.java:131) >>> at >>> org.apache.phoenix.jdbc.PhoenixEmbeddedDriver.connect(PhoenixEmbeddedDriver.java:112) >>> at java.sql.DriverManager.getConnection(DriverManager.java:571) >>> at java.sql.DriverManager.getConnection(DriverManager.java:187) >>> at org.apache.phoenix.util.PhoenixRuntime.main(PhoenixRuntime.java:197) >>> >>> java.sql.SQLException: ERROR 2006 (INT08): Incompatible jars detected >>> between client and server. Ensure that phoenix.jar is put on the classpath >>> of HBase in every region server: ERROR 1102 (XCL02): Cannot get all table >>> regions >>> at >>> org.apache.phoenix.exception.SQLExceptionCode$Factory$1.newException(SQLExceptionCode.java:309) >>> at >>> org.apache.phoenix.exception.SQLExceptionInfo.buildException(SQLExceptionInfo.java:133) >>> at >>> org.apache.phoenix.query.ConnectionQueryServicesImpl.checkClientServerCompatibility(ConnectionQueryServicesImpl.java:932) >>> at >>> org.apache.phoenix.query.ConnectionQueryServicesImpl.ensureTableCreated(ConnectionQueryServicesImpl.java:831) >>> at >>> org.apache.phoenix.query.ConnectionQueryServicesImpl.createTable(ConnectionQueryServicesImpl.java:1058) >>> at >>> org.apache.phoenix.schema.MetaDataClient.createTableInternal(MetaDataClient.java:1156) >>> at >>> org.apache.phoenix.schema.MetaDataClient.createTable(MetaDataClient.java:422) >>> at >>> org.apache.phoenix.compile.CreateTableCompiler$2.execute(CreateTableCompiler.java:183) >>> at >>> org.apache.phoenix.jdbc.PhoenixStatement.executeMutation(PhoenixStatement.java:226) >>> at >>> org.apache.phoenix.jdbc.PhoenixStatement.executeUpdate(PhoenixStatement.java:908) >>> at >>> org.apache.phoenix.query.ConnectionQueryServicesImpl.init(ConnectionQueryServicesImpl.java:1453) >>> at >>> org.apache.phoenix.jdbc.PhoenixDriver.getConnectionQueryServices(PhoenixDriver.java:131) >>> at >>> org.apache.phoenix.jdbc.PhoenixEmbeddedDriver.connect(PhoenixEmbeddedDriver.java:112) >>> at java.sql.DriverManager.getConnection(DriverManager.java:571) >>> at java.sql.DriverManager.getConnection(DriverManager.java:187) >>> at org.apache.phoenix.util.PhoenixRuntime.main(PhoenixRuntime.java:197) >>> Caused by: java.sql.SQLException: ERROR 1102 (XCL02): Cannot get all >>> table regions >>> at >>> org.apache.phoenix.exception.SQLExceptionCode$Factory$1.newException(SQLExceptionCode.java:309) >>> at >>> org.apache.phoenix.exception.SQLExceptionInfo.buildException(SQLExceptionInfo.java:133) >>> at >>> org.apache.phoenix.query.ConnectionQueryServicesImpl.getAllTableRegions(ConnectionQueryServicesImpl.java:425) >>> at >>> org.apache.phoenix.query.ConnectionQueryServicesImpl.checkClientServerCompatibility(ConnectionQueryServicesImpl.java:887) >>> ... 13 more >>> Caused by: org.apache.hadoop.hbase.client.NoServerForRegionException: No >>> server address listed in hbase:meta for region >>> SYSTEM.CATALOG,,1408684006641.0d1ea455127dd4af6a806574b1f42a91. containing >>> row >>> at >>> org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.locateRegionInMeta(ConnectionManager.java:1334) >>> at >>> org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.locateRegion(ConnectionManager.java:1128) >>> at >>> org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.relocateRegion(ConnectionManager.java:1097) >>> at >>> org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.relocateRegion(ConnectionManager.java:1084) >>> at >>> org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getRegionLocation(ConnectionManager.java:904) >>> at >>> org.apache.phoenix.query.ConnectionQueryServicesImpl.getAllTableRegions(ConnectionQueryServicesImpl.java:411) >>> ... 14 more >>> >>> -- >>> Regards, >>> Vikas Agarwal >>> 91 – 9928301411 >>> >>> InfoObjects, Inc. >>> Execution Matters >>> http://www.infoobjects.com >>> 2041 Mission College Boulevard, #280 >>> Santa Clara, CA 95054 >>> +1 (408) 988-2000 Work >>> +1 (408) 716-2726 Fax >> >> >> >> CONFIDENTIALITY NOTICE >> NOTICE: This message is intended for the use of the individual or entity >> to which it is addressed and may contain information that is confidential, >> privileged and exempt from disclosure under applicable law. If the reader of >> this message is not the intended recipient, you are hereby notified that any >> printing, copying, dissemination, distribution, disclosure or forwarding of >> this communication is strictly prohibited. If you have received this >> communication in error, please contact the sender immediately and delete it >> from your system. Thank You. > > > > > -- > Regards, > Vikas Agarwal > 91 – 9928301411 > > InfoObjects, Inc. > Execution Matters > http://www.infoobjects.com > 2041 Mission College Boulevard, #280 > Santa Clara, CA 95054 > +1 (408) 988-2000 Work > +1 (408) 716-2726 Fax