Github user rja1 commented on the issue:
https://github.com/apache/zeppelin/pull/1157
Thanks for the tip @prabhjyotsingh and for your work, hive is working!
I'm running into an issue with jdbc(phoenix) now though, hoping you can
help. My interpreter config is listed below. Note that we don't have a
/hbase-secure dir in zookeeper, just hbase.
[zk: localhost:2181(CONNECTED) 1] ls /hbase
[replication, meta-region-server, rs, splitWAL, backup-masters, table-lock,
flush-table-proc, region-in-transition, online-snapshot, acl, master, running,
balancer, tokenauth, recovering-regions, draining, namespace, hbaseid, table]
Here's the notebook / error:
select * from USER_ACCOUNTS where USER_SEED = '1000'
Failed after attempts=1, exceptions:
Mon Jul 18 14:00:22 MDT 2016,
RpcRetryingCaller{globalStartTime=1468872022128, pause=100, retries=1},
org.apache.hadoop.hbase.MasterNotRunningException:
com.google.protobuf.ServiceException:
org.apache.hadoop.hbase.exceptions.ConnectionClosingException: Call to
name01.hadoop.test.company.com/10.4.59.25:60000 failed on local exception:
org.apache.hadoop.hbase.exceptions.ConnectionClosingException: Connection to
name01.hadoop.test.company.com/10.4.59.25:60000 is closing. Call id=0,
waitTime=11
class org.apache.phoenix.exception.PhoenixIOException
org.apache.phoenix.util.ServerUtil.parseServerException(ServerUtil.java:111)
org.apache.phoenix.query.ConnectionQueryServicesImpl.ensureTableCreated(ConnectionQueryServicesImpl.java:1064)
org.apache.phoenix.query.ConnectionQueryServicesImpl.createTable(ConnectionQueryServicesImpl.java:1370)
org.apache.phoenix.schema.MetaDataClient.createTableInternal(MetaDataClient.java:2116)
org.apache.phoenix.schema.MetaDataClient.createTable(MetaDataClient.java:828)
org.apache.phoenix.compile.CreateTableCompiler$2.execute(CreateTableCompiler.java:183)
org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:338)
org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:326)
org.apache.phoenix.call.CallRunner.run(CallRunner.java:53)
org.apache.phoenix.jdbc.PhoenixStatement.executeMutation(PhoenixStatement.java:324)
org.apache.phoenix.jdbc.PhoenixStatement.executeUpdate(PhoenixStatement.java:1326)
org.apache.phoenix.query.ConnectionQueryServicesImpl$13.call(ConnectionQueryServicesImpl.java:2275)
org.apache.phoenix.query.ConnectionQueryServicesImpl$13.call(ConnectionQueryServicesImpl.java:2244)
org.apache.phoenix.util.PhoenixContextExecutor.call(PhoenixContextExecutor.java:78)
org.apache.phoenix.query.ConnectionQueryServicesImpl.init(ConnectionQueryServicesImpl.java:2244)
org.apache.phoenix.jdbc.PhoenixDriver.getConnectionQueryServices(PhoenixDriver.java:233)
org.apache.phoenix.jdbc.PhoenixEmbeddedDriver.createConnection(PhoenixEmbeddedDriver.java:135)
org.apache.phoenix.jdbc.PhoenixDriver.connect(PhoenixDriver.java:202)
java.sql.DriverManager.getConnection(DriverManager.java:664)
java.sql.DriverManager.getConnection(DriverManager.java:208)
org.apache.zeppelin.jdbc.JDBCInterpreter.getConnection(JDBCInterpreter.java:226)
org.apache.zeppelin.jdbc.JDBCInterpreter.getStatement(JDBCInterpreter.java:237)
org.apache.zeppelin.jdbc.JDBCInterpreter.executeSql(JDBCInterpreter.java:296)
org.apache.zeppelin.jdbc.JDBCInterpreter.interpret(JDBCInterpreter.java:402)
org.apache.zeppelin.interpreter.LazyOpenInterpreter.interpret(LazyOpenInterpreter.java:94)
org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:383)
org.apache.zeppelin.scheduler.Job.run(Job.java:176)
org.apache.zeppelin.scheduler.ParallelScheduler$JobRunner.run(ParallelScheduler.java:162)
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
java.util.concurrent.FutureTask.run(FutureTask.java:266)
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180)
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293)
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
java.lang.Thread.run(Thread.java:745)
Interpreter Config:
"2BRGRRCBW": {
"id": "2BRGRRCBW",
"name": "jdbc",
"group": "jdbc",
"properties": {
"phoenix.user": "zeppelin",
"hive.url":
"jdbc:hive2://cms01.hadoop.test.company.com:10000/default;principal\u003dhive/[email protected]",
"default.driver": "org.postgresql.Driver",
"phoenix.driver": "org.apache.phoenix.jdbc.PhoenixDriver",
"hive.user": "hive",
"psql.password": "",
"psql.user": "phoenixuser",
"psql.url": "jdbc:postgresql://localhost:5432/",
"default.user": "gpadmin",
"phoenix.hbase.client.retries.number": "1",
"phoenix.url": "jdbc:phoenix:cms01.hadoop.test.company.com:/hbase",
"tajo.url": "jdbc:tajo://localhost:26002/default",
"tajo.driver": "org.apache.tajo.jdbc.TajoDriver",
"psql.driver": "org.postgresql.Driver",
"default.password": "",
"zeppelin.interpreter.localRepo":
"/opt/zeppelin-ZEPPELIN-1146/local-repo/2BRGRRCBW",
"zeppelin.jdbc.auth.type": "KERBEROS",
"zeppelin.jdbc.concurrent.use": "true",
"hive.password": "",
"hive.driver": "org.apache.hive.jdbc.HiveDriver",
"zeppelin.jdbc.keytab.location":
"/opt/zeppelin-ZEPPELIN-1146/conf/zeppelin.keytab",
"common.max_count": "1000",
"phoenix.password": "",
"zeppelin.jdbc.principal": "[email protected]",
"zeppelin.jdbc.concurrent.max_connection": "10",
"default.url": "jdbc:postgresql://localhost:5432/"
},
"interpreterGroup": [
{
"name": "sql",
"class": "org.apache.zeppelin.jdbc.JDBCInterpreter"
}
],
"dependencies": [
{
"groupArtifactVersion":
"/opt/cloudera/CDH/jars/hive-jdbc-1.1.0-cdh5.5.2.jar",
"local": false
},
{
"groupArtifactVersion":
"/opt/cloudera/CDH/jars/hadoop-common-2.6.0-cdh5.5.2.jar",
"local": false
},
{
"groupArtifactVersion":
"/opt/cloudera/CDH/jars/hive-shims-0.23-1.1.0-cdh5.5.2.jar",
"local": false,
"exclusions": []
},
{
"groupArtifactVersion":
"/opt/cloudera/CDH/jars/hive-jdbc-1.1.0-cdh5.5.2-standalone.jar",
"local": false
},
{
"groupArtifactVersion":
"/opt/cloudera/CDH/jars/phoenix-core-4.7.0-cdh5.5.1.jar",
"local": false
},
{
"groupArtifactVersion":
"/opt/cloudera/CDH/jars/hbase-common-1.0.0-cdh5.5.2.jar",
"local": false
},
{
"groupArtifactVersion":
"/opt/cloudera/CDH/jars/hbase-client-1.0.0-cdh5.5.2.jar",
"local": false
},
{
"groupArtifactVersion":
"/opt/cloudera/CDH/jars/phoenix-4.7.0-cdh5.5.1-client.jar",
"local": false
}
],
"option": {
"remote": true,
"port": -1,
"perNoteSession": false,
"perNoteProcess": false,
"isExistingProcess": false
}
},
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at [email protected] or file a JIRA ticket
with INFRA.
---