Hi, The Hadoop version is Hadoop 3.1.0. Hbase is 1.2.6 and Phoenix is apache-phoenix-4.8.1-HBase-1.2-bin
In the past I had issues with Hbase 2 working with Hadoop 3.1 so I had to use Hbase 1.2.6. The individual components work fine. In other words I can do all operations on Hbase with Hadoop 3.1 and Phoenix. The issue I am facing is using both org.apache.phoenix.mapreduce.CsvBulkLoadTool and hbase.mapreduce.ImportTsv utilities. So I presume the issue may be to do with both these command line tools not working with Hadoop 3.1? Thanks Dr Mich Talebzadeh LinkedIn * https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw <https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw>* http://talebzadehmich.wordpress.com *Disclaimer:* Use it at your own risk. Any and all responsibility for any loss, damage or destruction of data or any other property which may arise from relying on this email's technical content is explicitly disclaimed. The author will in no case be liable for any monetary damages arising from such loss, damage or destruction. On Tue, 21 Aug 2018 at 00:48, Sergey Soldatov <sergey.solda...@gmail.com> wrote: > If I read it correctly you are trying to use Phoenix and HBase that were > built against Hadoop 2 with Hadoop 3. Is HBase was the only component you > have upgraded? > > Thanks, > Sergey > > On Mon, Aug 20, 2018 at 1:42 PM Mich Talebzadeh <mich.talebza...@gmail.com> > wrote: > >> Here you go >> >> 2018-08-20 18:29:47,248 INFO [main] zookeeper.ZooKeeper: Client >> environment:java.library.path=/home/hduser/hadoop-3.1.0/lib >> 2018-08-20 18:29:47,248 INFO [main] zookeeper.ZooKeeper: Client >> environment:java.io.tmpdir=/tmp >> 2018-08-20 18:29:47,248 INFO [main] zookeeper.ZooKeeper: Client >> environment:java.compiler=<NA> >> 2018-08-20 18:29:47,248 INFO [main] zookeeper.ZooKeeper: Client >> environment:os.name=Linux >> 2018-08-20 18:29:47,248 INFO [main] zookeeper.ZooKeeper: Client >> environment:os.arch=amd64 >> 2018-08-20 18:29:47,248 INFO [main] zookeeper.ZooKeeper: Client >> environment:os.version=3.10.0-862.3.2.el7.x86_64 >> 2018-08-20 18:29:47,248 INFO [main] zookeeper.ZooKeeper: Client >> environment:user.name=hduser >> 2018-08-20 18:29:47,248 INFO [main] zookeeper.ZooKeeper: Client >> environment:user.home=/home/hduser >> 2018-08-20 18:29:47,248 INFO [main] zookeeper.ZooKeeper: Client >> environment:user.dir=/data6/hduser/streaming_data/2018-08-20 >> 2018-08-20 18:29:47,249 INFO [main] zookeeper.ZooKeeper: Initiating >> client connection, connectString=rhes75:2181 sessionTimeout=90000 >> watcher=hconnection-0x493d44230x0, quorum=rhes75:2181, baseZNode=/hbase >> 2018-08-20 18:29:47,261 INFO [main-SendThread(rhes75:2181)] >> zookeeper.ClientCnxn: Opening socket connection to server rhes75/ >> 50.140.197.220:2181. Will not attempt to authenticate using SASL >> (unknown error) >> 2018-08-20 18:29:47,264 INFO [main-SendThread(rhes75:2181)] >> zookeeper.ClientCnxn: Socket connection established to rhes75/ >> 50.140.197.220:2181, initiating session >> 2018-08-20 18:29:47,281 INFO [main-SendThread(rhes75:2181)] >> zookeeper.ClientCnxn: Session establishment complete on server rhes75/ >> 50.140.197.220:2181, sessionid = 0x1002ea99eed0077, negotiated timeout = >> 40000 >> Exception in thread "main" java.sql.SQLException: ERROR 103 (08004): >> Unable to establish connection. >> at >> org.apache.phoenix.exception.SQLExceptionCode$Factory$1.newException(SQLExceptionCode.java:455) >> at >> org.apache.phoenix.exception.SQLExceptionInfo.buildException(SQLExceptionInfo.java:145) >> at >> org.apache.phoenix.query.ConnectionQueryServicesImpl.openConnection(ConnectionQueryServicesImpl.java:386) >> at >> org.apache.phoenix.query.ConnectionQueryServicesImpl.access$300(ConnectionQueryServicesImpl.java:222) >> at >> org.apache.phoenix.query.ConnectionQueryServicesImpl$13.call(ConnectionQueryServicesImpl.java:2318) >> at >> org.apache.phoenix.query.ConnectionQueryServicesImpl$13.call(ConnectionQueryServicesImpl.java:2294) >> at >> org.apache.phoenix.util.PhoenixContextExecutor.call(PhoenixContextExecutor.java:76) >> at >> org.apache.phoenix.query.ConnectionQueryServicesImpl.init(ConnectionQueryServicesImpl.java:2294) >> at >> org.apache.phoenix.jdbc.PhoenixDriver.getConnectionQueryServices(PhoenixDriver.java:232) >> at >> org.apache.phoenix.jdbc.PhoenixEmbeddedDriver.createConnection(PhoenixEmbeddedDriver.java:147) >> at >> org.apache.phoenix.jdbc.PhoenixDriver.connect(PhoenixDriver.java:202) >> at java.sql.DriverManager.getConnection(DriverManager.java:664) >> at java.sql.DriverManager.getConnection(DriverManager.java:208) >> at >> org.apache.phoenix.util.QueryUtil.getConnection(QueryUtil.java:340) >> at >> org.apache.phoenix.util.QueryUtil.getConnection(QueryUtil.java:332) >> at >> org.apache.phoenix.mapreduce.AbstractBulkLoadTool.loadData(AbstractBulkLoadTool.java:209) >> at >> org.apache.phoenix.mapreduce.AbstractBulkLoadTool.run(AbstractBulkLoadTool.java:183) >> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:76) >> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:90) >> at >> org.apache.phoenix.mapreduce.CsvBulkLoadTool.main(CsvBulkLoadTool.java:101) >> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >> at >> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) >> at >> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >> at java.lang.reflect.Method.invoke(Method.java:498) >> at org.apache.hadoop.util.RunJar.run(RunJar.java:308) >> at org.apache.hadoop.util.RunJar.main(RunJar.java:222) >> Caused by: java.io.IOException: >> java.lang.reflect.InvocationTargetException >> at >> org.apache.hadoop.hbase.client.ConnectionFactory.createConnection(ConnectionFactory.java:240) >> at >> org.apache.hadoop.hbase.client.ConnectionManager.createConnection(ConnectionManager.java:431) >> at >> org.apache.hadoop.hbase.client.ConnectionManager.createConnectionInternal(ConnectionManager.java:340) >> at >> org.apache.hadoop.hbase.client.HConnectionManager.createConnection(HConnectionManager.java:144) >> at >> org.apache.phoenix.query.HConnectionFactory$HConnectionFactoryImpl.createConnection(HConnectionFactory.java:47) >> at >> org.apache.phoenix.query.ConnectionQueryServicesImpl.openConnection(ConnectionQueryServicesImpl.java:383) >> ... 23 more >> Caused by: java.lang.reflect.InvocationTargetException >> at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native >> Method) >> at >> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) >> at >> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) >> at java.lang.reflect.Constructor.newInstance(Constructor.java:423) >> at >> org.apache.hadoop.hbase.client.ConnectionFactory.createConnection(ConnectionFactory.java:238) >> ... 28 more >> Caused by: java.lang.IllegalAccessError: class >> org.apache.hadoop.hdfs.web.HftpFileSystem cannot access its superinterface >> org.apache.hadoop.hdfs.web.TokenAspect$TokenManagementDelegator >> at java.lang.ClassLoader.defineClass1(Native Method) >> at java.lang.ClassLoader.defineClass(ClassLoader.java:763) >> at >> java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142) >> at java.net.URLClassLoader.defineClass(URLClassLoader.java:467) >> at java.net.URLClassLoader.access$100(URLClassLoader.java:73) >> at java.net.URLClassLoader$1.run(URLClassLoader.java:368) >> at java.net.URLClassLoader$1.run(URLClassLoader.java:362) >> at java.security.AccessController.doPrivileged(Native Method) >> at java.net.URLClassLoader.findClass(URLClassLoader.java:361) >> at java.lang.ClassLoader.loadClass(ClassLoader.java:424) >> at java.lang.ClassLoader.loadClass(ClassLoader.java:357) >> at java.lang.Class.forName0(Native Method) >> at java.lang.Class.forName(Class.java:348) >> at >> java.util.ServiceLoader$LazyIterator.nextService(ServiceLoader.java:370) >> at >> java.util.ServiceLoader$LazyIterator.next(ServiceLoader.java:404) >> at java.util.ServiceLoader$1.next(ServiceLoader.java:480) >> at >> org.apache.hadoop.fs.FileSystem.loadFileSystems(FileSystem.java:3268) >> at >> org.apache.hadoop.fs.FileSystem.getFileSystemClass(FileSystem.java:3313) >> at >> org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:3352) >> at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:124) >> at >> org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:3403) >> at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:3371) >> at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:477) >> at org.apache.hadoop.fs.Path.getFileSystem(Path.java:361) >> at >> org.apache.hadoop.hbase.util.DynamicClassLoader.initTempDir(DynamicClassLoader.java:120) >> at >> org.apache.hadoop.hbase.util.DynamicClassLoader.<init>(DynamicClassLoader.java:98) >> at >> org.apache.hadoop.hbase.protobuf.ProtobufUtil.<clinit>(ProtobufUtil.java:242) >> at org.apache.hadoop.hbase.ClusterId.parseFrom(ClusterId.java:64) >> at >> org.apache.hadoop.hbase.zookeeper.ZKClusterId.readClusterIdZNode(ZKClusterId.java:75) >> at >> org.apache.hadoop.hbase.client.ZooKeeperRegistry.getClusterId(ZooKeeperRegistry.java:105) >> at >> org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.retrieveClusterId(ConnectionManager.java:905) >> at >> org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.<init>(ConnectionManager.java:648) >> ... 33 more >> >> Thanks >> >> Dr Mich Talebzadeh >> >> >> >> LinkedIn * >> https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw >> <https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw>* >> >> >> >> http://talebzadehmich.wordpress.com >> >> >> *Disclaimer:* Use it at your own risk. Any and all responsibility for >> any loss, damage or destruction of data or any other property which may >> arise from relying on this email's technical content is explicitly >> disclaimed. The author will in no case be liable for any monetary damages >> arising from such loss, damage or destruction. >> >> >> >> >> On Mon, 20 Aug 2018 at 21:24, Josh Elser <els...@apache.org> wrote: >> >>> (-cc user@hbase, +bcc user@hbase) >>> >>> How about the rest of the stacktrace? You didn't share the cause. >>> >>> On 8/20/18 1:35 PM, Mich Talebzadeh wrote: >>> > >>> > This was working fine before my Hbase upgrade to 1.2.6 >>> > >>> > I have Hbase version 1.2.6 and Phoenix >>> > version apache-phoenix-4.8.1-HBase-1.2-bin >>> > >>> > This command bulkloading into Hbase through phoenix failsnow fails >>> > >>> > >>> HADOOP_CLASSPATH=${HOME}/jars/hbase-protocol-1.2.6.jar:${HBASE_HOME}/conf >>> hadoop >>> > jar ${HBASE_HOME}/lib/phoenix-4.8.1-HBase-1.2-client.jar >>> > org.apache.phoenix.mapreduce.CsvBulkLoadTool --table ${TABLE_NAME} >>> > --input >>> hdfs://rhes75:9000/${REFINED_HBASE_SUB_DIR}/${FILE_NAME}_${dir}.txt >>> > >>> > hadoop jar >>> > /data6/hduser/hbase-1.2.6/lib/phoenix-4.8.1-HBase-1.2-client.jar >>> > org.apache.phoenix.mapreduce.CsvBulkLoadTool --table >>> > MARKETDATAHBASEBATCH --input >>> > >>> hdfs://rhes75:9000//data/prices/2018-08-20_refined/populate_Phoenix_table_MARKETDATAHBASEBATCH_2018-08-20.txt >>> > + >>> > >>> HADOOP_CLASSPATH=/home/hduser/jars/hbase-protocol-1.2.6.jar:/data6/hduser/hbase-1.2.6/conf >>> > >>> > >>> > With the following error >>> > >>> > 2018-08-20 18:29:47,248 INFO [main] zookeeper.ZooKeeper: Client >>> > environment:java.library.path=/home/hduser/hadoop-3.1.0/lib >>> > 2018-08-20 18:29:47,248 INFO [main] zookeeper.ZooKeeper: Client >>> > environment:java.io.tmpdir=/tmp >>> > 2018-08-20 18:29:47,248 INFO [main] zookeeper.ZooKeeper: Client >>> > environment:java.compiler=<NA> >>> > 2018-08-20 18:29:47,248 INFO [main] zookeeper.ZooKeeper: Client >>> > environment:os.name <http://os.name>=Linux >>> > 2018-08-20 18:29:47,248 INFO [main] zookeeper.ZooKeeper: Client >>> > environment:os.arch=amd64 >>> > 2018-08-20 18:29:47,248 INFO [main] zookeeper.ZooKeeper: Client >>> > environment:os.version=3.10.0-862.3.2.el7.x86_64 >>> > 2018-08-20 18:29:47,248 INFO [main] zookeeper.ZooKeeper: Client >>> > environment:user.name <http://user.name>=hduser >>> > 2018-08-20 18:29:47,248 INFO [main] zookeeper.ZooKeeper: Client >>> > environment:user.home=/home/hduser >>> > 2018-08-20 18:29:47,248 INFO [main] zookeeper.ZooKeeper: Client >>> > environment:user.dir=/data6/hduser/streaming_data/2018-08-20 >>> > 2018-08-20 18:29:47,249 INFO [main] zookeeper.ZooKeeper: Initiating >>> > client connection, connectString=rhes75:2181 sessionTimeout=90000 >>> > watcher=hconnection-0x493d44230x0, quorum=rhes75:2181, baseZNode=/hbase >>> > 2018-08-20 18:29:47,261 INFO [main-SendThread(rhes75:2181)] >>> > zookeeper.ClientCnxn: Opening socket connection to server >>> > rhes75/50.140.197.220:2181 <http://50.140.197.220:2181>. Will not >>> > attempt to authenticate using SASL (unknown error) >>> > 2018-08-20 18:29:47,264 INFO [main-SendThread(rhes75:2181)] >>> > zookeeper.ClientCnxn: Socket connection established to >>> > rhes75/50.140.197.220:2181 <http://50.140.197.220:2181>, initiating >>> session >>> > 2018-08-20 18:29:47,281 INFO [main-SendThread(rhes75:2181)] >>> > zookeeper.ClientCnxn: Session establishment complete on server >>> > rhes75/50.140.197.220:2181 <http://50.140.197.220:2181>, sessionid = >>> > 0x1002ea99eed0077, negotiated timeout = 40000 >>> > Exception in thread "main" java.sql.SQLException: ERROR 103 (08004): >>> > Unable to establish connection. >>> > at >>> > >>> org.apache.phoenix.exception.SQLExceptionCode$Factory$1.newException(SQLExceptionCode.java:455) >>> > >>> > Any thoughts? >>> > >>> > Thanks >>> > >>> > Dr Mich Talebzadeh >>> > >>> > LinkedIn >>> > / >>> https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw/ >>> > >>> > http://talebzadehmich.wordpress.com >>> > >>> > >>> > *Disclaimer:* Use it at your own risk.Any and all responsibility for >>> any >>> > loss, damage or destruction of data or any other property which may >>> > arise from relying on this email's technical content is explicitly >>> > disclaimed. The author will in no case be liable for any monetary >>> > damages arising from such loss, damage or destruction. >>> > >>> >>