If I read it correctly you are trying to use Phoenix and HBase that were
built against Hadoop 2 with Hadoop 3. Is HBase was the only component you
have upgraded?

Thanks,
Sergey

On Mon, Aug 20, 2018 at 1:42 PM Mich Talebzadeh <mich.talebza...@gmail.com>
wrote:

> Here you go
>
> 2018-08-20 18:29:47,248 INFO  [main] zookeeper.ZooKeeper: Client
> environment:java.library.path=/home/hduser/hadoop-3.1.0/lib
> 2018-08-20 18:29:47,248 INFO  [main] zookeeper.ZooKeeper: Client
> environment:java.io.tmpdir=/tmp
> 2018-08-20 18:29:47,248 INFO  [main] zookeeper.ZooKeeper: Client
> environment:java.compiler=<NA>
> 2018-08-20 18:29:47,248 INFO  [main] zookeeper.ZooKeeper: Client
> environment:os.name=Linux
> 2018-08-20 18:29:47,248 INFO  [main] zookeeper.ZooKeeper: Client
> environment:os.arch=amd64
> 2018-08-20 18:29:47,248 INFO  [main] zookeeper.ZooKeeper: Client
> environment:os.version=3.10.0-862.3.2.el7.x86_64
> 2018-08-20 18:29:47,248 INFO  [main] zookeeper.ZooKeeper: Client
> environment:user.name=hduser
> 2018-08-20 18:29:47,248 INFO  [main] zookeeper.ZooKeeper: Client
> environment:user.home=/home/hduser
> 2018-08-20 18:29:47,248 INFO  [main] zookeeper.ZooKeeper: Client
> environment:user.dir=/data6/hduser/streaming_data/2018-08-20
> 2018-08-20 18:29:47,249 INFO  [main] zookeeper.ZooKeeper: Initiating
> client connection, connectString=rhes75:2181 sessionTimeout=90000
> watcher=hconnection-0x493d44230x0, quorum=rhes75:2181, baseZNode=/hbase
> 2018-08-20 18:29:47,261 INFO  [main-SendThread(rhes75:2181)]
> zookeeper.ClientCnxn: Opening socket connection to server rhes75/
> 50.140.197.220:2181. Will not attempt to authenticate using SASL (unknown
> error)
> 2018-08-20 18:29:47,264 INFO  [main-SendThread(rhes75:2181)]
> zookeeper.ClientCnxn: Socket connection established to rhes75/
> 50.140.197.220:2181, initiating session
> 2018-08-20 18:29:47,281 INFO  [main-SendThread(rhes75:2181)]
> zookeeper.ClientCnxn: Session establishment complete on server rhes75/
> 50.140.197.220:2181, sessionid = 0x1002ea99eed0077, negotiated timeout =
> 40000
> Exception in thread "main" java.sql.SQLException: ERROR 103 (08004):
> Unable to establish connection.
>         at
> org.apache.phoenix.exception.SQLExceptionCode$Factory$1.newException(SQLExceptionCode.java:455)
>         at
> org.apache.phoenix.exception.SQLExceptionInfo.buildException(SQLExceptionInfo.java:145)
>         at
> org.apache.phoenix.query.ConnectionQueryServicesImpl.openConnection(ConnectionQueryServicesImpl.java:386)
>         at
> org.apache.phoenix.query.ConnectionQueryServicesImpl.access$300(ConnectionQueryServicesImpl.java:222)
>         at
> org.apache.phoenix.query.ConnectionQueryServicesImpl$13.call(ConnectionQueryServicesImpl.java:2318)
>         at
> org.apache.phoenix.query.ConnectionQueryServicesImpl$13.call(ConnectionQueryServicesImpl.java:2294)
>         at
> org.apache.phoenix.util.PhoenixContextExecutor.call(PhoenixContextExecutor.java:76)
>         at
> org.apache.phoenix.query.ConnectionQueryServicesImpl.init(ConnectionQueryServicesImpl.java:2294)
>         at
> org.apache.phoenix.jdbc.PhoenixDriver.getConnectionQueryServices(PhoenixDriver.java:232)
>         at
> org.apache.phoenix.jdbc.PhoenixEmbeddedDriver.createConnection(PhoenixEmbeddedDriver.java:147)
>         at
> org.apache.phoenix.jdbc.PhoenixDriver.connect(PhoenixDriver.java:202)
>         at java.sql.DriverManager.getConnection(DriverManager.java:664)
>         at java.sql.DriverManager.getConnection(DriverManager.java:208)
>         at
> org.apache.phoenix.util.QueryUtil.getConnection(QueryUtil.java:340)
>         at
> org.apache.phoenix.util.QueryUtil.getConnection(QueryUtil.java:332)
>         at
> org.apache.phoenix.mapreduce.AbstractBulkLoadTool.loadData(AbstractBulkLoadTool.java:209)
>         at
> org.apache.phoenix.mapreduce.AbstractBulkLoadTool.run(AbstractBulkLoadTool.java:183)
>         at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:76)
>         at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:90)
>         at
> org.apache.phoenix.mapreduce.CsvBulkLoadTool.main(CsvBulkLoadTool.java:101)
>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>         at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>         at java.lang.reflect.Method.invoke(Method.java:498)
>         at org.apache.hadoop.util.RunJar.run(RunJar.java:308)
>         at org.apache.hadoop.util.RunJar.main(RunJar.java:222)
> Caused by: java.io.IOException: java.lang.reflect.InvocationTargetException
>         at
> org.apache.hadoop.hbase.client.ConnectionFactory.createConnection(ConnectionFactory.java:240)
>         at
> org.apache.hadoop.hbase.client.ConnectionManager.createConnection(ConnectionManager.java:431)
>         at
> org.apache.hadoop.hbase.client.ConnectionManager.createConnectionInternal(ConnectionManager.java:340)
>         at
> org.apache.hadoop.hbase.client.HConnectionManager.createConnection(HConnectionManager.java:144)
>         at
> org.apache.phoenix.query.HConnectionFactory$HConnectionFactoryImpl.createConnection(HConnectionFactory.java:47)
>         at
> org.apache.phoenix.query.ConnectionQueryServicesImpl.openConnection(ConnectionQueryServicesImpl.java:383)
>         ... 23 more
> Caused by: java.lang.reflect.InvocationTargetException
>         at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native
> Method)
>         at
> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
>         at
> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
>         at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
>         at
> org.apache.hadoop.hbase.client.ConnectionFactory.createConnection(ConnectionFactory.java:238)
>         ... 28 more
> Caused by: java.lang.IllegalAccessError: class
> org.apache.hadoop.hdfs.web.HftpFileSystem cannot access its superinterface
> org.apache.hadoop.hdfs.web.TokenAspect$TokenManagementDelegator
>         at java.lang.ClassLoader.defineClass1(Native Method)
>         at java.lang.ClassLoader.defineClass(ClassLoader.java:763)
>         at
> java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142)
>         at java.net.URLClassLoader.defineClass(URLClassLoader.java:467)
>         at java.net.URLClassLoader.access$100(URLClassLoader.java:73)
>         at java.net.URLClassLoader$1.run(URLClassLoader.java:368)
>         at java.net.URLClassLoader$1.run(URLClassLoader.java:362)
>         at java.security.AccessController.doPrivileged(Native Method)
>         at java.net.URLClassLoader.findClass(URLClassLoader.java:361)
>         at java.lang.ClassLoader.loadClass(ClassLoader.java:424)
>         at java.lang.ClassLoader.loadClass(ClassLoader.java:357)
>         at java.lang.Class.forName0(Native Method)
>         at java.lang.Class.forName(Class.java:348)
>         at
> java.util.ServiceLoader$LazyIterator.nextService(ServiceLoader.java:370)
>         at
> java.util.ServiceLoader$LazyIterator.next(ServiceLoader.java:404)
>         at java.util.ServiceLoader$1.next(ServiceLoader.java:480)
>         at
> org.apache.hadoop.fs.FileSystem.loadFileSystems(FileSystem.java:3268)
>         at
> org.apache.hadoop.fs.FileSystem.getFileSystemClass(FileSystem.java:3313)
>         at
> org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:3352)
>         at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:124)
>         at
> org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:3403)
>         at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:3371)
>         at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:477)
>         at org.apache.hadoop.fs.Path.getFileSystem(Path.java:361)
>         at
> org.apache.hadoop.hbase.util.DynamicClassLoader.initTempDir(DynamicClassLoader.java:120)
>         at
> org.apache.hadoop.hbase.util.DynamicClassLoader.<init>(DynamicClassLoader.java:98)
>         at
> org.apache.hadoop.hbase.protobuf.ProtobufUtil.<clinit>(ProtobufUtil.java:242)
>         at org.apache.hadoop.hbase.ClusterId.parseFrom(ClusterId.java:64)
>         at
> org.apache.hadoop.hbase.zookeeper.ZKClusterId.readClusterIdZNode(ZKClusterId.java:75)
>         at
> org.apache.hadoop.hbase.client.ZooKeeperRegistry.getClusterId(ZooKeeperRegistry.java:105)
>         at
> org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.retrieveClusterId(ConnectionManager.java:905)
>         at
> org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.<init>(ConnectionManager.java:648)
>         ... 33 more
>
> Thanks
>
> Dr Mich Talebzadeh
>
>
>
> LinkedIn * 
> https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw
> <https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw>*
>
>
>
> http://talebzadehmich.wordpress.com
>
>
> *Disclaimer:* Use it at your own risk. Any and all responsibility for any
> loss, damage or destruction of data or any other property which may arise
> from relying on this email's technical content is explicitly disclaimed.
> The author will in no case be liable for any monetary damages arising from
> such loss, damage or destruction.
>
>
>
>
> On Mon, 20 Aug 2018 at 21:24, Josh Elser <els...@apache.org> wrote:
>
>> (-cc user@hbase, +bcc user@hbase)
>>
>> How about the rest of the stacktrace? You didn't share the cause.
>>
>> On 8/20/18 1:35 PM, Mich Talebzadeh wrote:
>> >
>> > This was working fine before my Hbase upgrade to 1.2.6
>> >
>> > I have Hbase version 1.2.6 and Phoenix
>> > version apache-phoenix-4.8.1-HBase-1.2-bin
>> >
>> > This command bulkloading into Hbase through phoenix failsnow fails
>> >
>> >
>> HADOOP_CLASSPATH=${HOME}/jars/hbase-protocol-1.2.6.jar:${HBASE_HOME}/conf
>> hadoop
>> > jar ${HBASE_HOME}/lib/phoenix-4.8.1-HBase-1.2-client.jar
>> > org.apache.phoenix.mapreduce.CsvBulkLoadTool --table ${TABLE_NAME}
>> > --input
>> hdfs://rhes75:9000/${REFINED_HBASE_SUB_DIR}/${FILE_NAME}_${dir}.txt
>> >
>> > hadoop jar
>> > /data6/hduser/hbase-1.2.6/lib/phoenix-4.8.1-HBase-1.2-client.jar
>> > org.apache.phoenix.mapreduce.CsvBulkLoadTool --table
>> > MARKETDATAHBASEBATCH --input
>> >
>> hdfs://rhes75:9000//data/prices/2018-08-20_refined/populate_Phoenix_table_MARKETDATAHBASEBATCH_2018-08-20.txt
>> > +
>> >
>> HADOOP_CLASSPATH=/home/hduser/jars/hbase-protocol-1.2.6.jar:/data6/hduser/hbase-1.2.6/conf
>> >
>> >
>> > With the following error
>> >
>> > 2018-08-20 18:29:47,248 INFO  [main] zookeeper.ZooKeeper: Client
>> > environment:java.library.path=/home/hduser/hadoop-3.1.0/lib
>> > 2018-08-20 18:29:47,248 INFO  [main] zookeeper.ZooKeeper: Client
>> > environment:java.io.tmpdir=/tmp
>> > 2018-08-20 18:29:47,248 INFO  [main] zookeeper.ZooKeeper: Client
>> > environment:java.compiler=<NA>
>> > 2018-08-20 18:29:47,248 INFO  [main] zookeeper.ZooKeeper: Client
>> > environment:os.name <http://os.name>=Linux
>> > 2018-08-20 18:29:47,248 INFO  [main] zookeeper.ZooKeeper: Client
>> > environment:os.arch=amd64
>> > 2018-08-20 18:29:47,248 INFO  [main] zookeeper.ZooKeeper: Client
>> > environment:os.version=3.10.0-862.3.2.el7.x86_64
>> > 2018-08-20 18:29:47,248 INFO  [main] zookeeper.ZooKeeper: Client
>> > environment:user.name <http://user.name>=hduser
>> > 2018-08-20 18:29:47,248 INFO  [main] zookeeper.ZooKeeper: Client
>> > environment:user.home=/home/hduser
>> > 2018-08-20 18:29:47,248 INFO  [main] zookeeper.ZooKeeper: Client
>> > environment:user.dir=/data6/hduser/streaming_data/2018-08-20
>> > 2018-08-20 18:29:47,249 INFO  [main] zookeeper.ZooKeeper: Initiating
>> > client connection, connectString=rhes75:2181 sessionTimeout=90000
>> > watcher=hconnection-0x493d44230x0, quorum=rhes75:2181, baseZNode=/hbase
>> > 2018-08-20 18:29:47,261 INFO  [main-SendThread(rhes75:2181)]
>> > zookeeper.ClientCnxn: Opening socket connection to server
>> > rhes75/50.140.197.220:2181 <http://50.140.197.220:2181>. Will not
>> > attempt to authenticate using SASL (unknown error)
>> > 2018-08-20 18:29:47,264 INFO  [main-SendThread(rhes75:2181)]
>> > zookeeper.ClientCnxn: Socket connection established to
>> > rhes75/50.140.197.220:2181 <http://50.140.197.220:2181>, initiating
>> session
>> > 2018-08-20 18:29:47,281 INFO  [main-SendThread(rhes75:2181)]
>> > zookeeper.ClientCnxn: Session establishment complete on server
>> > rhes75/50.140.197.220:2181 <http://50.140.197.220:2181>, sessionid =
>> > 0x1002ea99eed0077, negotiated timeout = 40000
>> > Exception in thread "main" java.sql.SQLException: ERROR 103 (08004):
>> > Unable to establish connection.
>> >          at
>> >
>> org.apache.phoenix.exception.SQLExceptionCode$Factory$1.newException(SQLExceptionCode.java:455)
>> >
>> > Any thoughts?
>> >
>> > Thanks
>> >
>> > Dr Mich Talebzadeh
>> >
>> > LinkedIn
>> > /
>> https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw/
>> >
>> > http://talebzadehmich.wordpress.com
>> >
>> >
>> > *Disclaimer:* Use it at your own risk.Any and all responsibility for
>> any
>> > loss, damage or destruction of data or any other property which may
>> > arise from relying on this email's technical content is explicitly
>> > disclaimed. The author will in no case be liable for any monetary
>> > damages arising from such loss, damage or destruction.
>> >
>>
>

Reply via email to