If I read it correctly you are trying to use Phoenix and HBase that were built against Hadoop 2 with Hadoop 3. Is HBase was the only component you have upgraded?
Thanks, Sergey On Mon, Aug 20, 2018 at 1:42 PM Mich Talebzadeh <mich.talebza...@gmail.com> wrote: > Here you go > > 2018-08-20 18:29:47,248 INFO [main] zookeeper.ZooKeeper: Client > environment:java.library.path=/home/hduser/hadoop-3.1.0/lib > 2018-08-20 18:29:47,248 INFO [main] zookeeper.ZooKeeper: Client > environment:java.io.tmpdir=/tmp > 2018-08-20 18:29:47,248 INFO [main] zookeeper.ZooKeeper: Client > environment:java.compiler=<NA> > 2018-08-20 18:29:47,248 INFO [main] zookeeper.ZooKeeper: Client > environment:os.name=Linux > 2018-08-20 18:29:47,248 INFO [main] zookeeper.ZooKeeper: Client > environment:os.arch=amd64 > 2018-08-20 18:29:47,248 INFO [main] zookeeper.ZooKeeper: Client > environment:os.version=3.10.0-862.3.2.el7.x86_64 > 2018-08-20 18:29:47,248 INFO [main] zookeeper.ZooKeeper: Client > environment:user.name=hduser > 2018-08-20 18:29:47,248 INFO [main] zookeeper.ZooKeeper: Client > environment:user.home=/home/hduser > 2018-08-20 18:29:47,248 INFO [main] zookeeper.ZooKeeper: Client > environment:user.dir=/data6/hduser/streaming_data/2018-08-20 > 2018-08-20 18:29:47,249 INFO [main] zookeeper.ZooKeeper: Initiating > client connection, connectString=rhes75:2181 sessionTimeout=90000 > watcher=hconnection-0x493d44230x0, quorum=rhes75:2181, baseZNode=/hbase > 2018-08-20 18:29:47,261 INFO [main-SendThread(rhes75:2181)] > zookeeper.ClientCnxn: Opening socket connection to server rhes75/ > 50.140.197.220:2181. Will not attempt to authenticate using SASL (unknown > error) > 2018-08-20 18:29:47,264 INFO [main-SendThread(rhes75:2181)] > zookeeper.ClientCnxn: Socket connection established to rhes75/ > 50.140.197.220:2181, initiating session > 2018-08-20 18:29:47,281 INFO [main-SendThread(rhes75:2181)] > zookeeper.ClientCnxn: Session establishment complete on server rhes75/ > 50.140.197.220:2181, sessionid = 0x1002ea99eed0077, negotiated timeout = > 40000 > Exception in thread "main" java.sql.SQLException: ERROR 103 (08004): > Unable to establish connection. > at > org.apache.phoenix.exception.SQLExceptionCode$Factory$1.newException(SQLExceptionCode.java:455) > at > org.apache.phoenix.exception.SQLExceptionInfo.buildException(SQLExceptionInfo.java:145) > at > org.apache.phoenix.query.ConnectionQueryServicesImpl.openConnection(ConnectionQueryServicesImpl.java:386) > at > org.apache.phoenix.query.ConnectionQueryServicesImpl.access$300(ConnectionQueryServicesImpl.java:222) > at > org.apache.phoenix.query.ConnectionQueryServicesImpl$13.call(ConnectionQueryServicesImpl.java:2318) > at > org.apache.phoenix.query.ConnectionQueryServicesImpl$13.call(ConnectionQueryServicesImpl.java:2294) > at > org.apache.phoenix.util.PhoenixContextExecutor.call(PhoenixContextExecutor.java:76) > at > org.apache.phoenix.query.ConnectionQueryServicesImpl.init(ConnectionQueryServicesImpl.java:2294) > at > org.apache.phoenix.jdbc.PhoenixDriver.getConnectionQueryServices(PhoenixDriver.java:232) > at > org.apache.phoenix.jdbc.PhoenixEmbeddedDriver.createConnection(PhoenixEmbeddedDriver.java:147) > at > org.apache.phoenix.jdbc.PhoenixDriver.connect(PhoenixDriver.java:202) > at java.sql.DriverManager.getConnection(DriverManager.java:664) > at java.sql.DriverManager.getConnection(DriverManager.java:208) > at > org.apache.phoenix.util.QueryUtil.getConnection(QueryUtil.java:340) > at > org.apache.phoenix.util.QueryUtil.getConnection(QueryUtil.java:332) > at > org.apache.phoenix.mapreduce.AbstractBulkLoadTool.loadData(AbstractBulkLoadTool.java:209) > at > org.apache.phoenix.mapreduce.AbstractBulkLoadTool.run(AbstractBulkLoadTool.java:183) > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:76) > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:90) > at > org.apache.phoenix.mapreduce.CsvBulkLoadTool.main(CsvBulkLoadTool.java:101) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:498) > at org.apache.hadoop.util.RunJar.run(RunJar.java:308) > at org.apache.hadoop.util.RunJar.main(RunJar.java:222) > Caused by: java.io.IOException: java.lang.reflect.InvocationTargetException > at > org.apache.hadoop.hbase.client.ConnectionFactory.createConnection(ConnectionFactory.java:240) > at > org.apache.hadoop.hbase.client.ConnectionManager.createConnection(ConnectionManager.java:431) > at > org.apache.hadoop.hbase.client.ConnectionManager.createConnectionInternal(ConnectionManager.java:340) > at > org.apache.hadoop.hbase.client.HConnectionManager.createConnection(HConnectionManager.java:144) > at > org.apache.phoenix.query.HConnectionFactory$HConnectionFactoryImpl.createConnection(HConnectionFactory.java:47) > at > org.apache.phoenix.query.ConnectionQueryServicesImpl.openConnection(ConnectionQueryServicesImpl.java:383) > ... 23 more > Caused by: java.lang.reflect.InvocationTargetException > at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native > Method) > at > sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) > at > sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) > at java.lang.reflect.Constructor.newInstance(Constructor.java:423) > at > org.apache.hadoop.hbase.client.ConnectionFactory.createConnection(ConnectionFactory.java:238) > ... 28 more > Caused by: java.lang.IllegalAccessError: class > org.apache.hadoop.hdfs.web.HftpFileSystem cannot access its superinterface > org.apache.hadoop.hdfs.web.TokenAspect$TokenManagementDelegator > at java.lang.ClassLoader.defineClass1(Native Method) > at java.lang.ClassLoader.defineClass(ClassLoader.java:763) > at > java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142) > at java.net.URLClassLoader.defineClass(URLClassLoader.java:467) > at java.net.URLClassLoader.access$100(URLClassLoader.java:73) > at java.net.URLClassLoader$1.run(URLClassLoader.java:368) > at java.net.URLClassLoader$1.run(URLClassLoader.java:362) > at java.security.AccessController.doPrivileged(Native Method) > at java.net.URLClassLoader.findClass(URLClassLoader.java:361) > at java.lang.ClassLoader.loadClass(ClassLoader.java:424) > at java.lang.ClassLoader.loadClass(ClassLoader.java:357) > at java.lang.Class.forName0(Native Method) > at java.lang.Class.forName(Class.java:348) > at > java.util.ServiceLoader$LazyIterator.nextService(ServiceLoader.java:370) > at > java.util.ServiceLoader$LazyIterator.next(ServiceLoader.java:404) > at java.util.ServiceLoader$1.next(ServiceLoader.java:480) > at > org.apache.hadoop.fs.FileSystem.loadFileSystems(FileSystem.java:3268) > at > org.apache.hadoop.fs.FileSystem.getFileSystemClass(FileSystem.java:3313) > at > org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:3352) > at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:124) > at > org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:3403) > at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:3371) > at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:477) > at org.apache.hadoop.fs.Path.getFileSystem(Path.java:361) > at > org.apache.hadoop.hbase.util.DynamicClassLoader.initTempDir(DynamicClassLoader.java:120) > at > org.apache.hadoop.hbase.util.DynamicClassLoader.<init>(DynamicClassLoader.java:98) > at > org.apache.hadoop.hbase.protobuf.ProtobufUtil.<clinit>(ProtobufUtil.java:242) > at org.apache.hadoop.hbase.ClusterId.parseFrom(ClusterId.java:64) > at > org.apache.hadoop.hbase.zookeeper.ZKClusterId.readClusterIdZNode(ZKClusterId.java:75) > at > org.apache.hadoop.hbase.client.ZooKeeperRegistry.getClusterId(ZooKeeperRegistry.java:105) > at > org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.retrieveClusterId(ConnectionManager.java:905) > at > org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.<init>(ConnectionManager.java:648) > ... 33 more > > Thanks > > Dr Mich Talebzadeh > > > > LinkedIn * > https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw > <https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw>* > > > > http://talebzadehmich.wordpress.com > > > *Disclaimer:* Use it at your own risk. Any and all responsibility for any > loss, damage or destruction of data or any other property which may arise > from relying on this email's technical content is explicitly disclaimed. > The author will in no case be liable for any monetary damages arising from > such loss, damage or destruction. > > > > > On Mon, 20 Aug 2018 at 21:24, Josh Elser <els...@apache.org> wrote: > >> (-cc user@hbase, +bcc user@hbase) >> >> How about the rest of the stacktrace? You didn't share the cause. >> >> On 8/20/18 1:35 PM, Mich Talebzadeh wrote: >> > >> > This was working fine before my Hbase upgrade to 1.2.6 >> > >> > I have Hbase version 1.2.6 and Phoenix >> > version apache-phoenix-4.8.1-HBase-1.2-bin >> > >> > This command bulkloading into Hbase through phoenix failsnow fails >> > >> > >> HADOOP_CLASSPATH=${HOME}/jars/hbase-protocol-1.2.6.jar:${HBASE_HOME}/conf >> hadoop >> > jar ${HBASE_HOME}/lib/phoenix-4.8.1-HBase-1.2-client.jar >> > org.apache.phoenix.mapreduce.CsvBulkLoadTool --table ${TABLE_NAME} >> > --input >> hdfs://rhes75:9000/${REFINED_HBASE_SUB_DIR}/${FILE_NAME}_${dir}.txt >> > >> > hadoop jar >> > /data6/hduser/hbase-1.2.6/lib/phoenix-4.8.1-HBase-1.2-client.jar >> > org.apache.phoenix.mapreduce.CsvBulkLoadTool --table >> > MARKETDATAHBASEBATCH --input >> > >> hdfs://rhes75:9000//data/prices/2018-08-20_refined/populate_Phoenix_table_MARKETDATAHBASEBATCH_2018-08-20.txt >> > + >> > >> HADOOP_CLASSPATH=/home/hduser/jars/hbase-protocol-1.2.6.jar:/data6/hduser/hbase-1.2.6/conf >> > >> > >> > With the following error >> > >> > 2018-08-20 18:29:47,248 INFO [main] zookeeper.ZooKeeper: Client >> > environment:java.library.path=/home/hduser/hadoop-3.1.0/lib >> > 2018-08-20 18:29:47,248 INFO [main] zookeeper.ZooKeeper: Client >> > environment:java.io.tmpdir=/tmp >> > 2018-08-20 18:29:47,248 INFO [main] zookeeper.ZooKeeper: Client >> > environment:java.compiler=<NA> >> > 2018-08-20 18:29:47,248 INFO [main] zookeeper.ZooKeeper: Client >> > environment:os.name <http://os.name>=Linux >> > 2018-08-20 18:29:47,248 INFO [main] zookeeper.ZooKeeper: Client >> > environment:os.arch=amd64 >> > 2018-08-20 18:29:47,248 INFO [main] zookeeper.ZooKeeper: Client >> > environment:os.version=3.10.0-862.3.2.el7.x86_64 >> > 2018-08-20 18:29:47,248 INFO [main] zookeeper.ZooKeeper: Client >> > environment:user.name <http://user.name>=hduser >> > 2018-08-20 18:29:47,248 INFO [main] zookeeper.ZooKeeper: Client >> > environment:user.home=/home/hduser >> > 2018-08-20 18:29:47,248 INFO [main] zookeeper.ZooKeeper: Client >> > environment:user.dir=/data6/hduser/streaming_data/2018-08-20 >> > 2018-08-20 18:29:47,249 INFO [main] zookeeper.ZooKeeper: Initiating >> > client connection, connectString=rhes75:2181 sessionTimeout=90000 >> > watcher=hconnection-0x493d44230x0, quorum=rhes75:2181, baseZNode=/hbase >> > 2018-08-20 18:29:47,261 INFO [main-SendThread(rhes75:2181)] >> > zookeeper.ClientCnxn: Opening socket connection to server >> > rhes75/50.140.197.220:2181 <http://50.140.197.220:2181>. Will not >> > attempt to authenticate using SASL (unknown error) >> > 2018-08-20 18:29:47,264 INFO [main-SendThread(rhes75:2181)] >> > zookeeper.ClientCnxn: Socket connection established to >> > rhes75/50.140.197.220:2181 <http://50.140.197.220:2181>, initiating >> session >> > 2018-08-20 18:29:47,281 INFO [main-SendThread(rhes75:2181)] >> > zookeeper.ClientCnxn: Session establishment complete on server >> > rhes75/50.140.197.220:2181 <http://50.140.197.220:2181>, sessionid = >> > 0x1002ea99eed0077, negotiated timeout = 40000 >> > Exception in thread "main" java.sql.SQLException: ERROR 103 (08004): >> > Unable to establish connection. >> > at >> > >> org.apache.phoenix.exception.SQLExceptionCode$Factory$1.newException(SQLExceptionCode.java:455) >> > >> > Any thoughts? >> > >> > Thanks >> > >> > Dr Mich Talebzadeh >> > >> > LinkedIn >> > / >> https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw/ >> > >> > http://talebzadehmich.wordpress.com >> > >> > >> > *Disclaimer:* Use it at your own risk.Any and all responsibility for >> any >> > loss, damage or destruction of data or any other property which may >> > arise from relying on this email's technical content is explicitly >> > disclaimed. The author will in no case be liable for any monetary >> > damages arising from such loss, damage or destruction. >> > >> >