Thanks James,

I did. Unfortunately, Hortonworks only supports software that they provide in 
their stack, and they only provide 4.2.2 for our version of the HDP stack. Even 
if we upgrade to the latest HDP, we would only get support for Phoenix 4.4

FWIW, I am also unable to connect Aqua Data Studio to Phoenix using the new 
client jar (phoenix-4.6.0-HBase-0.98-client.jar

When I try, I get the stack trace below.

However, I am still able to connect using the old 4.2.2 jar.


java.sql.SQLException: ERROR 103 (08004): Unable to establish connection.
                at 
org.apache.phoenix.exception.SQLExceptionCode$Factory$1.newException(SQLExceptionCode.java:396)
                at 
org.apache.phoenix.exception.SQLExceptionInfo.buildException(SQLExceptionInfo.java:145)
                at 
org.apache.phoenix.query.ConnectionQueryServicesImpl.openConnection(ConnectionQueryServicesImpl.java:298)
                at 
org.apache.phoenix.query.ConnectionQueryServicesImpl.access$300(ConnectionQueryServicesImpl.java:179)
                at 
org.apache.phoenix.query.ConnectionQueryServicesImpl$12.call(ConnectionQueryServicesImpl.java:1919)
                at 
org.apache.phoenix.query.ConnectionQueryServicesImpl$12.call(ConnectionQueryServicesImpl.java:1898)
                at 
org.apache.phoenix.util.PhoenixContextExecutor.call(PhoenixContextExecutor.java:78)
                at 
org.apache.phoenix.query.ConnectionQueryServicesImpl.init(ConnectionQueryServicesImpl.java:1898)
                at 
org.apache.phoenix.jdbc.PhoenixDriver.getConnectionQueryServices(PhoenixDriver.java:180)
                at 
org.apache.phoenix.jdbc.PhoenixEmbeddedDriver.connect(PhoenixEmbeddedDriver.java:132)
                at 
org.apache.phoenix.jdbc.PhoenixDriver.connect(PhoenixDriver.java:151)
                at \\...\\ .\\हिñçêČάй語简�?한\\.Jᠻꐎꎂᢋ 9.KX(Unknown Source)
                at \\...\\ .\\हिñçêČάй語简�?한\\.Jᠻꐎꎂᢋ 9.au(Unknown Source)
                at \\...\\ .\\हिñçêČάй語简�?한\\.Jᠻꐎꎂᢋ 9.getConnection(Unknown 
Source)
                at \\...\\ .\\हिñçêČάй語简�?한\\.Jᠻꐎꎂᢋ 9.getConnection(Unknown 
Source)
                at 
com.aquafold.datastudio.mainframe.dialog.connection.diagnostics.yꑅꀱꏓᜪ 
import.dv(Unknown Source)
                at 
com.aquafold.datastudio.mainframe.dialog.connection.diagnostics.yꑅꀱꏓᜪ 
import.b(Unknown Source)
                at \\...\\ .\\हिñçêČάй語简�?한\\.bᡲꐢꐟꄦ 5 5.d(Unknown Source)
                at \\...\\ .\\हिñçêČάй語简�?한\\.bᡲꐢꐟꄦ 5 5.b(Unknown Source)
                at 
com.aquafold.datastudio.mainframe.dialog.connection.diagnostics.PingDialog$3.runTest(Unknown
 Source)
                at 
com.aquafold.datastudio.mainframe.dialog.connection.diagnostics.PingDialog$2.run(Unknown
 Source)
Caused by: java.io.IOException: java.lang.reflect.InvocationTargetException
                at 
org.apache.hadoop.hbase.client.ConnectionManager.createConnection(ConnectionManager.java:426)
                at 
org.apache.hadoop.hbase.client.ConnectionManager.createConnectionInternal(ConnectionManager.java:319)
                at 
org.apache.hadoop.hbase.client.HConnectionManager.createConnection(HConnectionManager.java:292)
                at 
org.apache.phoenix.query.HConnectionFactory$HConnectionFactoryImpl.createConnection(HConnectionFactory.java:47)
                at 
org.apache.phoenix.query.ConnectionQueryServicesImpl.openConnection(ConnectionQueryServicesImpl.java:296)
                ... 18 more



From: James Taylor [mailto:jamestay...@apache.org]
Sent: Wednesday, December 09, 2015 2:31 PM
To: user
Subject: Re: CsvBulkUpload not working after upgrade to 4.6

Zack,
Have you asked Hortonworks through your support channel? This sounds like an 
issue related to the HDP version you have - you need to confirm with them that 
upgrading to Phoenix 4.6.0 will work (and if there are any extra steps you need 
to take).

Thanks,
James



On Wed, Dec 9, 2015 at 10:41 AM, Riesland, Zack 
<zack.riesl...@sensus.com<mailto:zack.riesl...@sensus.com>> wrote:
Thanks Samarth,

I’m running hbase 0.98.4.2.2.8.0-3150 and phoenix 4.6.0-HBase-0.98

The hbase stuff is there via the HDP 2.2.8 install. It worked before upgrading 
to 4.6.

From: Samarth Jain [mailto:sama...@apache.org<mailto:sama...@apache.org>]
Sent: Wednesday, December 09, 2015 1:29 PM
To: user@phoenix.apache.org<mailto:user@phoenix.apache.org>
Subject: Re: CsvBulkUpload not working after upgrade to 4.6

Zack,

What version of HBase are you running? And which version of Phoenix 
(specifically 4.6-0.98 version or 4.6-1.x version)? FWIW, I don't see the 
MetaRegionTracker.java file in HBase branches 1.x and master. Maybe you don't 
have the right hbase-client jar in place?

- Samarth

On Wed, Dec 9, 2015 at 4:30 AM, Riesland, Zack 
<zack.riesl...@sensus.com<mailto:zack.riesl...@sensus.com>> wrote:
This morning I tried running the same operation from a data node as well as a 
name node, where phoenix 4.2 is completely gone, and I get the exact same error.



From: Riesland, Zack
Sent: Tuesday, December 08, 2015 8:42 PM
To: user@phoenix.apache.org<mailto:user@phoenix.apache.org>
Subject: CsvBulkUpload not working after upgrade to 4.6

I upgraded our cluster from 4.2.2 to 4.6.

After a few hiccups, everything seems to be working: I can connect and interact 
with the DB using Aqua Studio. My web stuff that queries Phoenix works, using 
the new client jar. My java code to connect and interact with the DB works, 
using the new client jar, etc.

With one exception: Csv Bulk Upload does not work with the new client jar – 
only with the old one (4.2.2.blah).

On my edge node, where I run this from, I upgraded phoenix using the same 
script. /usr/hdp/current/phoenix-client now points to a folder full of 4.6 
stuff. Permissions all seem to be correct. However, the command below fails 
with the error below.

If I replace the reference to the (4.6) phoenix-client.jar and point explicitly 
to the old 4.2 client jar, it still works.

Any ideas or suggestions?

Thanks!



HADOOP_CLASSPATH=/usr/hdp/current/hbase-master/conf/:/usr/hdp/current/hbase-master/lib/hbase-protocol.jar
 hadoop jar /usr/hdp/current/phoenix-client/phoenix-client.jar 
org.apache.phoenix.mapreduce.CsvBulkLoadTool -Dfs.permissions.umask-mode=000 
--z <server info> --table <table> --input <input>

15/12/08 15:37:01 INFO zookeeper.ZooKeeper: Initiating client connection, 
connectString=<connect string> sessionTimeout=120000 
watcher=hconnection-0x7728c6760x0, quorum=<stuff>
15/12/08 15:37:01 INFO zookeeper.ClientCnxn: Opening socket connection to .... 
Will not attempt to authenticate using SASL (unknown error)
15/12/08 15:37:01 INFO zookeeper.ClientCnxn: Socket connection established to 
..., initiating session
15/12/08 15:37:01 INFO zookeeper.ClientCnxn: Session establishment complete on 
server..., sessionid = 0x25110076f7fbf25, negotiated timeout = 40000
15/12/08 15:37:01 INFO client.HConnectionManager$HConnectionImplementation: 
Closing master protocol: MasterService
15/12/08 15:37:01 INFO client.HConnectionManager$HConnectionImplementation: 
Closing zookeeper sessionid=0x25110076f7fbf25
15/12/08 15:37:01 INFO zookeeper.ZooKeeper: Session: 0x25110076f7fbf25 closed
15/12/08 15:37:01 INFO zookeeper.ClientCnxn: EventThread shut down
15/12/08 15:37:01 INFO client.HConnectionManager$HConnectionImplementation: 
Closing zookeeper sessionid=0x35102a35baf2c0c
15/12/08 15:37:01 INFO zookeeper.ZooKeeper: Session: 0x35102a35baf2c0c closed
15/12/08 15:37:01 INFO zookeeper.ClientCnxn: EventThread shut down
Exception in thread "main" java.sql.SQLException: ERROR 2006 (INT08): 
Incompatible jars detected between client and server. Ensure that phoenix.jar 
is put on the classpath of HBase in every region server: 
org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos$MetaRegionServer.hasState()Z
                at 
org.apache.phoenix.exception.SQLExceptionCode$Factory$1.newException(SQLExceptionCode.java:396)
                at 
org.apache.phoenix.exception.SQLExceptionInfo.buildException(SQLExceptionInfo.java:145)
                at 
org.apache.phoenix.query.ConnectionQueryServicesImpl.checkClientServerCompatibility(ConnectionQueryServicesImpl.java:1000)
                at 
org.apache.phoenix.query.ConnectionQueryServicesImpl.ensureTableCreated(ConnectionQueryServicesImpl.java:879)
                at 
org.apache.phoenix.query.ConnectionQueryServicesImpl.createTable(ConnectionQueryServicesImpl.java:1225)
                at 
org.apache.phoenix.query.DelegateConnectionQueryServices.createTable(DelegateConnectionQueryServices.java:113)
                at 
org.apache.phoenix.schema.MetaDataClient.createTableInternal(MetaDataClient.java:2013)
                at 
org.apache.phoenix.schema.MetaDataClient.createTable(MetaDataClient.java:785)
                at 
org.apache.phoenix.compile.CreateTableCompiler$2.execute(CreateTableCompiler.java:186)
                at 
org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:319)
                at 
org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:311)
                at org.apache.phoenix.call.CallRunner.run(CallRunner.java:53)
                at 
org.apache.phoenix.jdbc.PhoenixStatement.executeMutation(PhoenixStatement.java:309)
                at 
org.apache.phoenix.jdbc.PhoenixStatement.executeUpdate(PhoenixStatement.java:1368)
                at 
org.apache.phoenix.query.ConnectionQueryServicesImpl$12.call(ConnectionQueryServicesImpl.java:1929)
                at 
org.apache.phoenix.query.ConnectionQueryServicesImpl$12.call(ConnectionQueryServicesImpl.java:1898)
                at 
org.apache.phoenix.util.PhoenixContextExecutor.call(PhoenixContextExecutor.java:78)
                at 
org.apache.phoenix.query.ConnectionQueryServicesImpl.init(ConnectionQueryServicesImpl.java:1898)
                at 
org.apache.phoenix.jdbc.PhoenixDriver.getConnectionQueryServices(PhoenixDriver.java:180)
                at 
org.apache.phoenix.jdbc.PhoenixEmbeddedDriver.connect(PhoenixEmbeddedDriver.java:132)
                at 
org.apache.phoenix.jdbc.PhoenixDriver.connect(PhoenixDriver.java:151)
                at java.sql.DriverManager.getConnection(DriverManager.java:571)
                at java.sql.DriverManager.getConnection(DriverManager.java:187)
                at 
org.apache.phoenix.util.QueryUtil.getConnection(QueryUtil.java:301)
                at 
org.apache.phoenix.util.QueryUtil.getConnection(QueryUtil.java:292)
                at 
org.apache.phoenix.mapreduce.CsvBulkLoadTool.loadData(CsvBulkLoadTool.java:211)
                at 
org.apache.phoenix.mapreduce.CsvBulkLoadTool.run(CsvBulkLoadTool.java:184)
                at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
                at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84)
                at 
org.apache.phoenix.mapreduce.CsvBulkLoadTool.main(CsvBulkLoadTool.java:99)
                at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
                at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
                at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
                at java.lang.reflect.Method.invoke(Method.java:606)
                at org.apache.hadoop.util.RunJar.run(RunJar.java:221)
                at org.apache.hadoop.util.RunJar.main(RunJar.java:136)
Caused by: java.lang.NoSuchMethodError: 
org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos$MetaRegionServer.hasState()Z
                at 
org.apache.hadoop.hbase.zookeeper.MetaRegionTracker.getMetaRegionState(MetaRegionTracker.java:219)
                at 
org.apache.hadoop.hbase.zookeeper.MetaRegionTracker.blockUntilAvailable(MetaRegionTracker.java:204)
                at 
org.apache.hadoop.hbase.client.ZooKeeperRegistry.getMetaRegionLocation(ZooKeeperRegistry.java:58)
                at 
org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.locateRegion(HConnectionManager.java:1157)
                at 
org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.locateRegionInMeta(HConnectionManager.java:1249)
                at 
org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.locateRegion(HConnectionManager.java:1160)
                at 
org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.locateRegion(HConnectionManager.java:1117)
                at 
org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.getRegionLocation(HConnectionManager.java:958)
                at 
org.apache.phoenix.query.ConnectionQueryServicesImpl.getAllTableRegions(ConnectionQueryServicesImpl.java:439)
                at 
org.apache.phoenix.query.ConnectionQueryServicesImpl.checkClientServerCompatibility(ConnectionQueryServicesImpl.java:953)
                ... 33 more


Reply via email to