Hi Gihan, I didn't use any tool to explorer the h2 database. just checked the published statistics in APIM publisher GUI and as you mentioned AUTOSERVER property is true by default. Any other clue to spot the error?
Thanks On Wed, Nov 5, 2014 at 1:58 PM, Gihan Anuruddha <[email protected]> wrote: > This can happen if you are using any tool to explore h2 database. Make > sure you add AUTO_SERVER=TRUE part at the end of connection url. > > <url>jdbc:h2:<BAM_HOME > >/repository/database/APIMGTSTATS_DB;AUTO_SERVER=TRUE</url> > > Regards, > Gihan > > On Wed, Nov 5, 2014 at 1:27 PM, Lakshman Udayakantha <[email protected]> > wrote: > >> I am working on this jira[1]. As per the offline discussion with Amani, I >> have to break the hive script to individual scripts to address follow >> analytic logics >> >> org_wso2_apimgt_statistics_destination >> >> org_wso2_apimgt_statistics_request >> >> org_wso2_apimgt_statistics_response >> >> org_wso2_apimgt_statistics_fault >> >> I have attached the analytic toolbox in which include four individual >> hive scripts to address above analytic logics here[2]. When I deployed this >> new toolbox on BAM, It posts some errors into the server logs. I pasted the >> server log part that contain these errors. Anyway I can notice published >> statistics on APIM dashboard. Anyone has any idea why I am getting these >> errors? >> >> [2014-11-05 12:46:15,955] ERROR >> {org.wso2.carbon.hadoop.hive.jdbc.storage.db.DBManager} - Failed to get >> connection >> >> org.h2.jdbc.JdbcSQLException: Connection is broken: "null" [90067-140] >> >> at org.h2.message.DbException.getJdbcSQLException(DbException.java:327) >> >> at org.h2.message.DbException.get(DbException.java:156) >> >> at org.h2.engine.SessionRemote.connectServer(SessionRemote.java:331) >> >> at >> org.h2.engine.SessionRemote.connectEmbeddedOrServer(SessionRemote.java:253) >> >> at org.h2.engine.SessionRemote.createSession(SessionRemote.java:219) >> >> at org.h2.jdbc.JdbcConnection.<init>(JdbcConnection.java:111) >> >> at org.h2.jdbc.JdbcConnection.<init>(JdbcConnection.java:95) >> >> at org.h2.Driver.connect(Driver.java:73) >> >> at java.sql.DriverManager.getConnection(DriverManager.java:582) >> >> at java.sql.DriverManager.getConnection(DriverManager.java:185) >> >> at >> org.wso2.carbon.hadoop.hive.jdbc.storage.db.DBManager.createConnection(DBManager.java:73) >> >> at >> org.wso2.carbon.hadoop.hive.jdbc.storage.db.DBManager.createConnection(DBManager.java:85) >> >> at >> org.wso2.carbon.hadoop.hive.jdbc.storage.JDBCDataOutputFormat.getHiveRecordWriter(JDBCDataOutputFormat.java:48) >> >> at >> org.apache.hadoop.hive.ql.io.HiveFileFormatUtils.getRecordWriter(HiveFileFormatUtils.java:236) >> >> at >> org.apache.hadoop.hive.ql.io.HiveFileFormatUtils.getHiveRecordWriter(HiveFileFormatUtils.java:224) >> >> at >> org.apache.hadoop.hive.ql.exec.FileSinkOperator.createBucketFiles(FileSinkOperator.java:478) >> >> at >> org.apache.hadoop.hive.ql.exec.FileSinkOperator.processOp(FileSinkOperator.java:526) >> >> at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:467) >> >> at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:758) >> >> at >> org.apache.hadoop.hive.ql.exec.SelectOperator.processOp(SelectOperator.java:84) >> >> at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:467) >> >> at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:758) >> >> at >> org.apache.hadoop.hive.ql.exec.SelectOperator.processOp(SelectOperator.java:84) >> >> at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:467) >> >> at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:758) >> >> at >> org.apache.hadoop.hive.ql.exec.GroupByOperator.forward(GroupByOperator.java:964) >> >> at >> org.apache.hadoop.hive.ql.exec.GroupByOperator.processAggr(GroupByOperator.java:781) >> >> at >> org.apache.hadoop.hive.ql.exec.GroupByOperator.processOp(GroupByOperator.java:707) >> >> at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:467) >> >> at org.apache.hadoop.hive.ql.exec.ExecReducer.reduce(ExecReducer.java:248) >> >> at org.apache.hadoop.mapred.ReduceTask.runOldReducer(ReduceTask.java:518) >> >> at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:419) >> >> at >> org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:257) >> >> Caused by: java.net.UnknownHostException: null >> >> at java.net.Inet6AddressImpl.lookupAllHostAddr(Native Method) >> >> at java.net.InetAddress$1.lookupAllHostAddr(InetAddress.java:883) >> >> at java.net.InetAddress.getAddressFromNameService(InetAddress.java:1236) >> >> at java.net.InetAddress.getAllByName0(InetAddress.java:1187) >> >> at java.net.InetAddress.getAllByName(InetAddress.java:1117) >> >> at java.net.InetAddress.getAllByName(InetAddress.java:1053) >> >> at java.net.InetAddress.getByName(InetAddress.java:1003) >> >> at org.h2.util.NetUtils.createSocket(NetUtils.java:90) >> >> at org.h2.engine.SessionRemote.initTransfer(SessionRemote.java:96) >> >> at org.h2.engine.SessionRemote.connectServer(SessionRemote.java:327) >> >> ... 30 more >> >> [2014-11-05 12:46:15,963] FATAL {ExecReducer} - >> org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error while >> processing row (tag=0) >> {"key":{"_col0":"testAPI2:v1.0.0","_col1":"carbon.super","_col2":"/test/api2","_col3":"10.100.5.161","_col4":2014,"_col5":10,"_col6":31,"_col7":10,"_col8":14,"_col9":"2014-10-31 >> 10:14"},"value":{"_col0":{"count":1,"sum":583.0},"_col1":1},"alias":0} >> >> at org.apache.hadoop.hive.ql.exec.ExecReducer.reduce(ExecReducer.java:257) >> >> at org.apache.hadoop.mapred.ReduceTask.runOldReducer(ReduceTask.java:518) >> >> at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:419) >> >> at >> org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:257) >> >> Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: >> java.lang.NullPointerException >> >> at >> org.apache.hadoop.hive.ql.exec.GroupByOperator.processOp(GroupByOperator.java:720) >> >> at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:467) >> >> at org.apache.hadoop.hive.ql.exec.ExecReducer.reduce(ExecReducer.java:248) >> >> ... 3 more >> >> Caused by: java.lang.NullPointerException >> >> at >> org.wso2.carbon.hadoop.hive.jdbc.storage.db.DBOperation.isRowExisting(DBOperation.java:144) >> >> at >> org.wso2.carbon.hadoop.hive.jdbc.storage.db.DBOperation.writeToDB(DBOperation.java:59) >> >> at >> org.wso2.carbon.hadoop.hive.jdbc.storage.db.DBRecordWriter.write(DBRecordWriter.java:35) >> >> at >> org.apache.hadoop.hive.ql.exec.FileSinkOperator.processOp(FileSinkOperator.java:589) >> >> at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:467) >> >> at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:758) >> >> at >> org.apache.hadoop.hive.ql.exec.SelectOperator.processOp(SelectOperator.java:84) >> >> at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:467) >> >> at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:758) >> >> at >> org.apache.hadoop.hive.ql.exec.SelectOperator.processOp(SelectOperator.java:84) >> >> at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:467) >> >> at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:758) >> >> at >> org.apache.hadoop.hive.ql.exec.GroupByOperator.forward(GroupByOperator.java:964) >> >> at >> org.apache.hadoop.hive.ql.exec.GroupByOperator.processAggr(GroupByOperator.java:781) >> >> at >> org.apache.hadoop.hive.ql.exec.GroupByOperator.processOp(GroupByOperator.java:707) >> >> ... 5 more >> >> [2014-11-05 12:46:15,964] WARN {org.apache.hadoop.mapred.LocalJobRunner} >> - job_local_0001 >> >> java.lang.RuntimeException: >> org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error while >> processing row (tag=0) >> {"key":{"_col0":"testAPI2:v1.0.0","_col1":"carbon.super","_col2":"/test/api2","_col3":"10.100.5.161","_col4":2014,"_col5":10,"_col6":31,"_col7":10,"_col8":14,"_col9":"2014-10-31 >> 10:14"},"value":{"_col0":{"count":1,"sum":583.0},"_col1":1},"alias":0} >> >> at org.apache.hadoop.hive.ql.exec.ExecReducer.reduce(ExecReducer.java:269) >> >> at org.apache.hadoop.mapred.ReduceTask.runOldReducer(ReduceTask.java:518) >> >> at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:419) >> >> at >> org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:257) >> >> Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime >> Error while processing row (tag=0) >> {"key":{"_col0":"testAPI2:v1.0.0","_col1":"carbon.super","_col2":"/test/api2","_col3":"10.100.5.161","_col4":2014,"_col5":10,"_col6":31,"_col7":10,"_col8":14,"_col9":"2014-10-31 >> 10:14"},"value":{"_col0":{"count":1,"sum":583.0},"_col1":1},"alias":0} >> >> at org.apache.hadoop.hive.ql.exec.ExecReducer.reduce(ExecReducer.java:257) >> >> ... 3 more >> >> Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: >> java.lang.NullPointerException >> >> at >> org.apache.hadoop.hive.ql.exec.GroupByOperator.processOp(GroupByOperator.java:720) >> >> at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:467) >> >> at org.apache.hadoop.hive.ql.exec.ExecReducer.reduce(ExecReducer.java:248) >> >> ... 3 more >> >> Caused by: java.lang.NullPointerException >> >> at >> org.wso2.carbon.hadoop.hive.jdbc.storage.db.DBOperation.isRowExisting(DBOperation.java:144) >> >> at >> org.wso2.carbon.hadoop.hive.jdbc.storage.db.DBOperation.writeToDB(DBOperation.java:59) >> >> at >> org.wso2.carbon.hadoop.hive.jdbc.storage.db.DBRecordWriter.write(DBRecordWriter.java:35) >> >> at >> org.apache.hadoop.hive.ql.exec.FileSinkOperator.processOp(FileSinkOperator.java:589) >> >> at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:467) >> >> at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:758) >> >> at >> org.apache.hadoop.hive.ql.exec.SelectOperator.processOp(SelectOperator.java:84) >> >> at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:467) >> >> at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:758) >> >> at >> org.apache.hadoop.hive.ql.exec.SelectOperator.processOp(SelectOperator.java:84) >> >> at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:467) >> >> at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:758) >> >> at >> org.apache.hadoop.hive.ql.exec.GroupByOperator.forward(GroupByOperator.java:964) >> >> at >> org.apache.hadoop.hive.ql.exec.GroupByOperator.processAggr(GroupByOperator.java:781) >> >> at >> org.apache.hadoop.hive.ql.exec.GroupByOperator.processOp(GroupByOperator.java:707) >> >> ... 5 more >> >> Ended Job = job_local_0001 with errors >> >> [2014-11-05 12:46:16,665] ERROR >> {org.apache.hadoop.hive.ql.exec.ExecDriver} - Ended Job = job_local_0001 >> with errors >> >> Error during job, obtaining debugging information... >> >> [2014-11-05 12:46:16,667] ERROR >> {org.apache.hadoop.hive.ql.exec.ExecDriver} - Error during job, obtaining >> debugging information... >> >> Execution failed with exit status: 2 >> >> [2014-11-05 12:46:16,827] ERROR {org.apache.hadoop.hive.ql.exec.Task} - >> Execution failed with exit status: 2 >> >> Obtaining error information >> >> [2014-11-05 12:46:16,827] ERROR {org.apache.hadoop.hive.ql.exec.Task} - >> Obtaining error information >> >> Task failed! >> >> Task ID: >> >> Stage-0 >> >> Logs: >> >> [2014-11-05 12:46:16,827] ERROR {org.apache.hadoop.hive.ql.exec.Task} - >> >> Task failed! >> >> Task ID: >> >> Stage-0 >> >> Logs: >> >> >> /Users/lakshman/Desktop/hiveBreaker/test2/wso2bam-2.5.0/repository/logs//wso2carbon.log >> >> [2014-11-05 12:46:16,827] ERROR {org.apache.hadoop.hive.ql.exec.Task} - >> /Users/lakshman/Desktop/hiveBreaker/test2/wso2bam-2.5.0/repository/logs//wso2carbon.log >> >> [2014-11-05 12:46:16,827] ERROR >> {org.apache.hadoop.hive.ql.exec.ExecDriver} - Execution failed with exit >> status: 2 >> >> FAILED: Execution Error, return code 2 from >> org.apache.hadoop.hive.ql.exec.MapRedTask >> >> [2014-11-05 12:46:16,828] ERROR {org.apache.hadoop.hive.ql.Driver} - >> FAILED: Execution Error, return code 2 from >> org.apache.hadoop.hive.ql.exec.MapRedTask >> >> [2014-11-05 12:46:16,832] ERROR >> {org.wso2.carbon.analytics.hive.impl.HiveExecutorServiceImpl} - Error >> while executing Hive script. >> >> Query returned non-zero code: 9, cause: FAILED: Execution Error, return >> code 2 from org.apache.hadoop.hive.ql.exec.MapRedTask >> >> java.sql.SQLException: Query returned non-zero code: 9, cause: FAILED: >> Execution Error, return code 2 from >> org.apache.hadoop.hive.ql.exec.MapRedTask >> >> at >> org.apache.hadoop.hive.jdbc.HiveStatement.executeQuery(HiveStatement.java:189) >> >> at >> org.wso2.carbon.analytics.hive.impl.HiveExecutorServiceImpl$ScriptCallable.executeHiveQuery(HiveExecutorServiceImpl.java:578) >> >> at >> org.wso2.carbon.analytics.hive.impl.HiveExecutorServiceImpl$ScriptCallable.call(HiveExecutorServiceImpl.java:286) >> >> at >> org.wso2.carbon.analytics.hive.impl.HiveExecutorServiceImpl$ScriptCallable.call(HiveExecutorServiceImpl.java:189) >> >> at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) >> >> at java.util.concurrent.FutureTask.run(FutureTask.java:138) >> >> at >> java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:895) >> >> at >> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:918) >> >> at java.lang.Thread.run(Thread.java:695) >> >> [2014-11-05 12:46:16,833] ERROR >> {org.wso2.carbon.analytics.hive.task.HiveScriptExecutorTask} - Error while >> executing script : am_response_stats_analyzer >> >> org.wso2.carbon.analytics.hive.exception.HiveExecutionException: Error >> while executing Hive script.Query returned non-zero code: 9, cause: FAILED: >> Execution Error, return code 2 from >> org.apache.hadoop.hive.ql.exec.MapRedTask >> >> at >> org.wso2.carbon.analytics.hive.impl.HiveExecutorServiceImpl.execute(HiveExecutorServiceImpl.java:115) >> >> at >> org.wso2.carbon.analytics.hive.task.HiveScriptExecutorTask.execute(HiveScriptExecutorTask.java:70) >> >> at >> org.wso2.carbon.ntask.core.impl.TaskQuartzJobAdapter.execute(TaskQuartzJobAdapter.java:67) >> >> at org.quartz.core.JobRunShell.run(JobRunShell.java:213) >> >> at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:439) >> >> at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) >> >> at java.util.concurrent.FutureTask.run(FutureTask.java:138) >> >> at >> java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:895) >> >> at >> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:918) >> >> at java.lang.Thread.run(Thread.java:695) >> >> 2014-11-05 12:46:17,620 null map = 100%, reduce = 100% >> >> >> [1] https://wso2.org/jira/browse/APIMANAGER-2992 >> >> [2] >> https://drive.google.com/file/d/0B9KDy4GJKr1vNFp0NEVwbnZhbjg/view?usp=sharing >> <https://drive.google.com/a/wso2.com/file/d/0Bx6Pq431GuEiSzBUVUxkNllnODQ/view?usp=sharing> >> >> -- >> Lakshman Udayakantha >> WSO2 Inc. www.wso2.com >> lean.enterprise.middleware >> Mobile: *0711241005 <0711241005>* >> >> >> _______________________________________________ >> Dev mailing list >> [email protected] >> http://wso2.org/cgi-bin/mailman/listinfo/dev >> >> > > > -- > W.G. Gihan Anuruddha > Senior Software Engineer | WSO2, Inc. > M: +94772272595 > -- Lakshman Udayakantha WSO2 Inc. www.wso2.com lean.enterprise.middleware Mobile: *0711241005*
_______________________________________________ Dev mailing list [email protected] http://wso2.org/cgi-bin/mailman/listinfo/dev
