This can happen if you are using any tool to explore h2 database. Make sure you add AUTO_SERVER=TRUE part at the end of connection url.
<url>jdbc:h2:<BAM_HOME >/repository/database/APIMGTSTATS_DB;AUTO_SERVER=TRUE</url> Regards, Gihan On Wed, Nov 5, 2014 at 1:27 PM, Lakshman Udayakantha <[email protected]> wrote: > I am working on this jira[1]. As per the offline discussion with Amani, I > have to break the hive script to individual scripts to address follow > analytic logics > > org_wso2_apimgt_statistics_destination > > org_wso2_apimgt_statistics_request > > org_wso2_apimgt_statistics_response > > org_wso2_apimgt_statistics_fault > > I have attached the analytic toolbox in which include four individual hive > scripts to address above analytic logics here[2]. When I deployed this new > toolbox on BAM, It posts some errors into the server logs. I pasted the > server log part that contain these errors. Anyway I can notice published > statistics on APIM dashboard. Anyone has any idea why I am getting these > errors? > > [2014-11-05 12:46:15,955] ERROR > {org.wso2.carbon.hadoop.hive.jdbc.storage.db.DBManager} - Failed to get > connection > > org.h2.jdbc.JdbcSQLException: Connection is broken: "null" [90067-140] > > at org.h2.message.DbException.getJdbcSQLException(DbException.java:327) > > at org.h2.message.DbException.get(DbException.java:156) > > at org.h2.engine.SessionRemote.connectServer(SessionRemote.java:331) > > at > org.h2.engine.SessionRemote.connectEmbeddedOrServer(SessionRemote.java:253) > > at org.h2.engine.SessionRemote.createSession(SessionRemote.java:219) > > at org.h2.jdbc.JdbcConnection.<init>(JdbcConnection.java:111) > > at org.h2.jdbc.JdbcConnection.<init>(JdbcConnection.java:95) > > at org.h2.Driver.connect(Driver.java:73) > > at java.sql.DriverManager.getConnection(DriverManager.java:582) > > at java.sql.DriverManager.getConnection(DriverManager.java:185) > > at > org.wso2.carbon.hadoop.hive.jdbc.storage.db.DBManager.createConnection(DBManager.java:73) > > at > org.wso2.carbon.hadoop.hive.jdbc.storage.db.DBManager.createConnection(DBManager.java:85) > > at > org.wso2.carbon.hadoop.hive.jdbc.storage.JDBCDataOutputFormat.getHiveRecordWriter(JDBCDataOutputFormat.java:48) > > at > org.apache.hadoop.hive.ql.io.HiveFileFormatUtils.getRecordWriter(HiveFileFormatUtils.java:236) > > at > org.apache.hadoop.hive.ql.io.HiveFileFormatUtils.getHiveRecordWriter(HiveFileFormatUtils.java:224) > > at > org.apache.hadoop.hive.ql.exec.FileSinkOperator.createBucketFiles(FileSinkOperator.java:478) > > at > org.apache.hadoop.hive.ql.exec.FileSinkOperator.processOp(FileSinkOperator.java:526) > > at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:467) > > at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:758) > > at > org.apache.hadoop.hive.ql.exec.SelectOperator.processOp(SelectOperator.java:84) > > at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:467) > > at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:758) > > at > org.apache.hadoop.hive.ql.exec.SelectOperator.processOp(SelectOperator.java:84) > > at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:467) > > at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:758) > > at > org.apache.hadoop.hive.ql.exec.GroupByOperator.forward(GroupByOperator.java:964) > > at > org.apache.hadoop.hive.ql.exec.GroupByOperator.processAggr(GroupByOperator.java:781) > > at > org.apache.hadoop.hive.ql.exec.GroupByOperator.processOp(GroupByOperator.java:707) > > at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:467) > > at org.apache.hadoop.hive.ql.exec.ExecReducer.reduce(ExecReducer.java:248) > > at org.apache.hadoop.mapred.ReduceTask.runOldReducer(ReduceTask.java:518) > > at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:419) > > at org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:257) > > Caused by: java.net.UnknownHostException: null > > at java.net.Inet6AddressImpl.lookupAllHostAddr(Native Method) > > at java.net.InetAddress$1.lookupAllHostAddr(InetAddress.java:883) > > at java.net.InetAddress.getAddressFromNameService(InetAddress.java:1236) > > at java.net.InetAddress.getAllByName0(InetAddress.java:1187) > > at java.net.InetAddress.getAllByName(InetAddress.java:1117) > > at java.net.InetAddress.getAllByName(InetAddress.java:1053) > > at java.net.InetAddress.getByName(InetAddress.java:1003) > > at org.h2.util.NetUtils.createSocket(NetUtils.java:90) > > at org.h2.engine.SessionRemote.initTransfer(SessionRemote.java:96) > > at org.h2.engine.SessionRemote.connectServer(SessionRemote.java:327) > > ... 30 more > > [2014-11-05 12:46:15,963] FATAL {ExecReducer} - > org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error while > processing row (tag=0) > {"key":{"_col0":"testAPI2:v1.0.0","_col1":"carbon.super","_col2":"/test/api2","_col3":"10.100.5.161","_col4":2014,"_col5":10,"_col6":31,"_col7":10,"_col8":14,"_col9":"2014-10-31 > 10:14"},"value":{"_col0":{"count":1,"sum":583.0},"_col1":1},"alias":0} > > at org.apache.hadoop.hive.ql.exec.ExecReducer.reduce(ExecReducer.java:257) > > at org.apache.hadoop.mapred.ReduceTask.runOldReducer(ReduceTask.java:518) > > at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:419) > > at org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:257) > > Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: > java.lang.NullPointerException > > at > org.apache.hadoop.hive.ql.exec.GroupByOperator.processOp(GroupByOperator.java:720) > > at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:467) > > at org.apache.hadoop.hive.ql.exec.ExecReducer.reduce(ExecReducer.java:248) > > ... 3 more > > Caused by: java.lang.NullPointerException > > at > org.wso2.carbon.hadoop.hive.jdbc.storage.db.DBOperation.isRowExisting(DBOperation.java:144) > > at > org.wso2.carbon.hadoop.hive.jdbc.storage.db.DBOperation.writeToDB(DBOperation.java:59) > > at > org.wso2.carbon.hadoop.hive.jdbc.storage.db.DBRecordWriter.write(DBRecordWriter.java:35) > > at > org.apache.hadoop.hive.ql.exec.FileSinkOperator.processOp(FileSinkOperator.java:589) > > at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:467) > > at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:758) > > at > org.apache.hadoop.hive.ql.exec.SelectOperator.processOp(SelectOperator.java:84) > > at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:467) > > at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:758) > > at > org.apache.hadoop.hive.ql.exec.SelectOperator.processOp(SelectOperator.java:84) > > at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:467) > > at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:758) > > at > org.apache.hadoop.hive.ql.exec.GroupByOperator.forward(GroupByOperator.java:964) > > at > org.apache.hadoop.hive.ql.exec.GroupByOperator.processAggr(GroupByOperator.java:781) > > at > org.apache.hadoop.hive.ql.exec.GroupByOperator.processOp(GroupByOperator.java:707) > > ... 5 more > > [2014-11-05 12:46:15,964] WARN {org.apache.hadoop.mapred.LocalJobRunner} > - job_local_0001 > > java.lang.RuntimeException: > org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error while > processing row (tag=0) > {"key":{"_col0":"testAPI2:v1.0.0","_col1":"carbon.super","_col2":"/test/api2","_col3":"10.100.5.161","_col4":2014,"_col5":10,"_col6":31,"_col7":10,"_col8":14,"_col9":"2014-10-31 > 10:14"},"value":{"_col0":{"count":1,"sum":583.0},"_col1":1},"alias":0} > > at org.apache.hadoop.hive.ql.exec.ExecReducer.reduce(ExecReducer.java:269) > > at org.apache.hadoop.mapred.ReduceTask.runOldReducer(ReduceTask.java:518) > > at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:419) > > at org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:257) > > Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime > Error while processing row (tag=0) > {"key":{"_col0":"testAPI2:v1.0.0","_col1":"carbon.super","_col2":"/test/api2","_col3":"10.100.5.161","_col4":2014,"_col5":10,"_col6":31,"_col7":10,"_col8":14,"_col9":"2014-10-31 > 10:14"},"value":{"_col0":{"count":1,"sum":583.0},"_col1":1},"alias":0} > > at org.apache.hadoop.hive.ql.exec.ExecReducer.reduce(ExecReducer.java:257) > > ... 3 more > > Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: > java.lang.NullPointerException > > at > org.apache.hadoop.hive.ql.exec.GroupByOperator.processOp(GroupByOperator.java:720) > > at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:467) > > at org.apache.hadoop.hive.ql.exec.ExecReducer.reduce(ExecReducer.java:248) > > ... 3 more > > Caused by: java.lang.NullPointerException > > at > org.wso2.carbon.hadoop.hive.jdbc.storage.db.DBOperation.isRowExisting(DBOperation.java:144) > > at > org.wso2.carbon.hadoop.hive.jdbc.storage.db.DBOperation.writeToDB(DBOperation.java:59) > > at > org.wso2.carbon.hadoop.hive.jdbc.storage.db.DBRecordWriter.write(DBRecordWriter.java:35) > > at > org.apache.hadoop.hive.ql.exec.FileSinkOperator.processOp(FileSinkOperator.java:589) > > at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:467) > > at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:758) > > at > org.apache.hadoop.hive.ql.exec.SelectOperator.processOp(SelectOperator.java:84) > > at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:467) > > at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:758) > > at > org.apache.hadoop.hive.ql.exec.SelectOperator.processOp(SelectOperator.java:84) > > at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:467) > > at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:758) > > at > org.apache.hadoop.hive.ql.exec.GroupByOperator.forward(GroupByOperator.java:964) > > at > org.apache.hadoop.hive.ql.exec.GroupByOperator.processAggr(GroupByOperator.java:781) > > at > org.apache.hadoop.hive.ql.exec.GroupByOperator.processOp(GroupByOperator.java:707) > > ... 5 more > > Ended Job = job_local_0001 with errors > > [2014-11-05 12:46:16,665] ERROR > {org.apache.hadoop.hive.ql.exec.ExecDriver} - Ended Job = job_local_0001 > with errors > > Error during job, obtaining debugging information... > > [2014-11-05 12:46:16,667] ERROR > {org.apache.hadoop.hive.ql.exec.ExecDriver} - Error during job, obtaining > debugging information... > > Execution failed with exit status: 2 > > [2014-11-05 12:46:16,827] ERROR {org.apache.hadoop.hive.ql.exec.Task} - > Execution failed with exit status: 2 > > Obtaining error information > > [2014-11-05 12:46:16,827] ERROR {org.apache.hadoop.hive.ql.exec.Task} - > Obtaining error information > > Task failed! > > Task ID: > > Stage-0 > > Logs: > > [2014-11-05 12:46:16,827] ERROR {org.apache.hadoop.hive.ql.exec.Task} - > > Task failed! > > Task ID: > > Stage-0 > > Logs: > > > /Users/lakshman/Desktop/hiveBreaker/test2/wso2bam-2.5.0/repository/logs//wso2carbon.log > > [2014-11-05 12:46:16,827] ERROR {org.apache.hadoop.hive.ql.exec.Task} - > /Users/lakshman/Desktop/hiveBreaker/test2/wso2bam-2.5.0/repository/logs//wso2carbon.log > > [2014-11-05 12:46:16,827] ERROR > {org.apache.hadoop.hive.ql.exec.ExecDriver} - Execution failed with exit > status: 2 > > FAILED: Execution Error, return code 2 from > org.apache.hadoop.hive.ql.exec.MapRedTask > > [2014-11-05 12:46:16,828] ERROR {org.apache.hadoop.hive.ql.Driver} - > FAILED: Execution Error, return code 2 from > org.apache.hadoop.hive.ql.exec.MapRedTask > > [2014-11-05 12:46:16,832] ERROR > {org.wso2.carbon.analytics.hive.impl.HiveExecutorServiceImpl} - Error > while executing Hive script. > > Query returned non-zero code: 9, cause: FAILED: Execution Error, return > code 2 from org.apache.hadoop.hive.ql.exec.MapRedTask > > java.sql.SQLException: Query returned non-zero code: 9, cause: FAILED: > Execution Error, return code 2 from > org.apache.hadoop.hive.ql.exec.MapRedTask > > at > org.apache.hadoop.hive.jdbc.HiveStatement.executeQuery(HiveStatement.java:189) > > at > org.wso2.carbon.analytics.hive.impl.HiveExecutorServiceImpl$ScriptCallable.executeHiveQuery(HiveExecutorServiceImpl.java:578) > > at > org.wso2.carbon.analytics.hive.impl.HiveExecutorServiceImpl$ScriptCallable.call(HiveExecutorServiceImpl.java:286) > > at > org.wso2.carbon.analytics.hive.impl.HiveExecutorServiceImpl$ScriptCallable.call(HiveExecutorServiceImpl.java:189) > > at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) > > at java.util.concurrent.FutureTask.run(FutureTask.java:138) > > at > java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:895) > > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:918) > > at java.lang.Thread.run(Thread.java:695) > > [2014-11-05 12:46:16,833] ERROR > {org.wso2.carbon.analytics.hive.task.HiveScriptExecutorTask} - Error while > executing script : am_response_stats_analyzer > > org.wso2.carbon.analytics.hive.exception.HiveExecutionException: Error > while executing Hive script.Query returned non-zero code: 9, cause: FAILED: > Execution Error, return code 2 from > org.apache.hadoop.hive.ql.exec.MapRedTask > > at > org.wso2.carbon.analytics.hive.impl.HiveExecutorServiceImpl.execute(HiveExecutorServiceImpl.java:115) > > at > org.wso2.carbon.analytics.hive.task.HiveScriptExecutorTask.execute(HiveScriptExecutorTask.java:70) > > at > org.wso2.carbon.ntask.core.impl.TaskQuartzJobAdapter.execute(TaskQuartzJobAdapter.java:67) > > at org.quartz.core.JobRunShell.run(JobRunShell.java:213) > > at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:439) > > at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) > > at java.util.concurrent.FutureTask.run(FutureTask.java:138) > > at > java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:895) > > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:918) > > at java.lang.Thread.run(Thread.java:695) > > 2014-11-05 12:46:17,620 null map = 100%, reduce = 100% > > > [1] https://wso2.org/jira/browse/APIMANAGER-2992 > > [2] > https://drive.google.com/file/d/0B9KDy4GJKr1vNFp0NEVwbnZhbjg/view?usp=sharing > <https://drive.google.com/a/wso2.com/file/d/0Bx6Pq431GuEiSzBUVUxkNllnODQ/view?usp=sharing> > > -- > Lakshman Udayakantha > WSO2 Inc. www.wso2.com > lean.enterprise.middleware > Mobile: *0711241005 <0711241005>* > > > _______________________________________________ > Dev mailing list > [email protected] > http://wso2.org/cgi-bin/mailman/listinfo/dev > > -- W.G. Gihan Anuruddha Senior Software Engineer | WSO2, Inc. M: +94772272595
_______________________________________________ Dev mailing list [email protected] http://wso2.org/cgi-bin/mailman/listinfo/dev
