I am working on this jira[1]. As per the offline discussion with Amani, I
have to break the hive script to individual scripts to address follow
analytic logics

org_wso2_apimgt_statistics_destination

org_wso2_apimgt_statistics_request

org_wso2_apimgt_statistics_response

org_wso2_apimgt_statistics_fault

I have attached the analytic toolbox in which include four individual hive
scripts to address above analytic logics here[2]. When I deployed this new
toolbox on BAM, It posts some errors into the server logs. I pasted the
server log part that contain these errors. Anyway I can notice published
statistics on APIM dashboard. Anyone has any idea why I am getting these
errors?

[2014-11-05 12:46:15,955] ERROR
{org.wso2.carbon.hadoop.hive.jdbc.storage.db.DBManager} -  Failed to get
connection

org.h2.jdbc.JdbcSQLException: Connection is broken: "null" [90067-140]

at org.h2.message.DbException.getJdbcSQLException(DbException.java:327)

at org.h2.message.DbException.get(DbException.java:156)

at org.h2.engine.SessionRemote.connectServer(SessionRemote.java:331)

at
org.h2.engine.SessionRemote.connectEmbeddedOrServer(SessionRemote.java:253)

at org.h2.engine.SessionRemote.createSession(SessionRemote.java:219)

at org.h2.jdbc.JdbcConnection.<init>(JdbcConnection.java:111)

at org.h2.jdbc.JdbcConnection.<init>(JdbcConnection.java:95)

at org.h2.Driver.connect(Driver.java:73)

at java.sql.DriverManager.getConnection(DriverManager.java:582)

at java.sql.DriverManager.getConnection(DriverManager.java:185)

at
org.wso2.carbon.hadoop.hive.jdbc.storage.db.DBManager.createConnection(DBManager.java:73)

at
org.wso2.carbon.hadoop.hive.jdbc.storage.db.DBManager.createConnection(DBManager.java:85)

at
org.wso2.carbon.hadoop.hive.jdbc.storage.JDBCDataOutputFormat.getHiveRecordWriter(JDBCDataOutputFormat.java:48)

at
org.apache.hadoop.hive.ql.io.HiveFileFormatUtils.getRecordWriter(HiveFileFormatUtils.java:236)

at
org.apache.hadoop.hive.ql.io.HiveFileFormatUtils.getHiveRecordWriter(HiveFileFormatUtils.java:224)

at
org.apache.hadoop.hive.ql.exec.FileSinkOperator.createBucketFiles(FileSinkOperator.java:478)

at
org.apache.hadoop.hive.ql.exec.FileSinkOperator.processOp(FileSinkOperator.java:526)

at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:467)

at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:758)

at
org.apache.hadoop.hive.ql.exec.SelectOperator.processOp(SelectOperator.java:84)

at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:467)

at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:758)

at
org.apache.hadoop.hive.ql.exec.SelectOperator.processOp(SelectOperator.java:84)

at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:467)

at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:758)

at
org.apache.hadoop.hive.ql.exec.GroupByOperator.forward(GroupByOperator.java:964)

at
org.apache.hadoop.hive.ql.exec.GroupByOperator.processAggr(GroupByOperator.java:781)

at
org.apache.hadoop.hive.ql.exec.GroupByOperator.processOp(GroupByOperator.java:707)

at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:467)

at org.apache.hadoop.hive.ql.exec.ExecReducer.reduce(ExecReducer.java:248)

at org.apache.hadoop.mapred.ReduceTask.runOldReducer(ReduceTask.java:518)

at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:419)

at org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:257)

Caused by: java.net.UnknownHostException: null

at java.net.Inet6AddressImpl.lookupAllHostAddr(Native Method)

at java.net.InetAddress$1.lookupAllHostAddr(InetAddress.java:883)

at java.net.InetAddress.getAddressFromNameService(InetAddress.java:1236)

at java.net.InetAddress.getAllByName0(InetAddress.java:1187)

at java.net.InetAddress.getAllByName(InetAddress.java:1117)

at java.net.InetAddress.getAllByName(InetAddress.java:1053)

at java.net.InetAddress.getByName(InetAddress.java:1003)

at org.h2.util.NetUtils.createSocket(NetUtils.java:90)

at org.h2.engine.SessionRemote.initTransfer(SessionRemote.java:96)

at org.h2.engine.SessionRemote.connectServer(SessionRemote.java:327)

... 30 more

[2014-11-05 12:46:15,963] FATAL {ExecReducer} -
org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error while
processing row (tag=0)
{"key":{"_col0":"testAPI2:v1.0.0","_col1":"carbon.super","_col2":"/test/api2","_col3":"10.100.5.161","_col4":2014,"_col5":10,"_col6":31,"_col7":10,"_col8":14,"_col9":"2014-10-31
10:14"},"value":{"_col0":{"count":1,"sum":583.0},"_col1":1},"alias":0}

at org.apache.hadoop.hive.ql.exec.ExecReducer.reduce(ExecReducer.java:257)

at org.apache.hadoop.mapred.ReduceTask.runOldReducer(ReduceTask.java:518)

at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:419)

at org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:257)

Caused by: org.apache.hadoop.hive.ql.metadata.HiveException:
java.lang.NullPointerException

at
org.apache.hadoop.hive.ql.exec.GroupByOperator.processOp(GroupByOperator.java:720)

at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:467)

at org.apache.hadoop.hive.ql.exec.ExecReducer.reduce(ExecReducer.java:248)

... 3 more

Caused by: java.lang.NullPointerException

at
org.wso2.carbon.hadoop.hive.jdbc.storage.db.DBOperation.isRowExisting(DBOperation.java:144)

at
org.wso2.carbon.hadoop.hive.jdbc.storage.db.DBOperation.writeToDB(DBOperation.java:59)

at
org.wso2.carbon.hadoop.hive.jdbc.storage.db.DBRecordWriter.write(DBRecordWriter.java:35)

at
org.apache.hadoop.hive.ql.exec.FileSinkOperator.processOp(FileSinkOperator.java:589)

at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:467)

at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:758)

at
org.apache.hadoop.hive.ql.exec.SelectOperator.processOp(SelectOperator.java:84)

at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:467)

at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:758)

at
org.apache.hadoop.hive.ql.exec.SelectOperator.processOp(SelectOperator.java:84)

at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:467)

at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:758)

at
org.apache.hadoop.hive.ql.exec.GroupByOperator.forward(GroupByOperator.java:964)

at
org.apache.hadoop.hive.ql.exec.GroupByOperator.processAggr(GroupByOperator.java:781)

at
org.apache.hadoop.hive.ql.exec.GroupByOperator.processOp(GroupByOperator.java:707)

... 5 more

[2014-11-05 12:46:15,964]  WARN {org.apache.hadoop.mapred.LocalJobRunner}
-  job_local_0001

java.lang.RuntimeException:
org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error while
processing row (tag=0)
{"key":{"_col0":"testAPI2:v1.0.0","_col1":"carbon.super","_col2":"/test/api2","_col3":"10.100.5.161","_col4":2014,"_col5":10,"_col6":31,"_col7":10,"_col8":14,"_col9":"2014-10-31
10:14"},"value":{"_col0":{"count":1,"sum":583.0},"_col1":1},"alias":0}

at org.apache.hadoop.hive.ql.exec.ExecReducer.reduce(ExecReducer.java:269)

at org.apache.hadoop.mapred.ReduceTask.runOldReducer(ReduceTask.java:518)

at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:419)

at org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:257)

Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime
Error while processing row (tag=0)
{"key":{"_col0":"testAPI2:v1.0.0","_col1":"carbon.super","_col2":"/test/api2","_col3":"10.100.5.161","_col4":2014,"_col5":10,"_col6":31,"_col7":10,"_col8":14,"_col9":"2014-10-31
10:14"},"value":{"_col0":{"count":1,"sum":583.0},"_col1":1},"alias":0}

at org.apache.hadoop.hive.ql.exec.ExecReducer.reduce(ExecReducer.java:257)

... 3 more

Caused by: org.apache.hadoop.hive.ql.metadata.HiveException:
java.lang.NullPointerException

at
org.apache.hadoop.hive.ql.exec.GroupByOperator.processOp(GroupByOperator.java:720)

at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:467)

at org.apache.hadoop.hive.ql.exec.ExecReducer.reduce(ExecReducer.java:248)

... 3 more

Caused by: java.lang.NullPointerException

at
org.wso2.carbon.hadoop.hive.jdbc.storage.db.DBOperation.isRowExisting(DBOperation.java:144)

at
org.wso2.carbon.hadoop.hive.jdbc.storage.db.DBOperation.writeToDB(DBOperation.java:59)

at
org.wso2.carbon.hadoop.hive.jdbc.storage.db.DBRecordWriter.write(DBRecordWriter.java:35)

at
org.apache.hadoop.hive.ql.exec.FileSinkOperator.processOp(FileSinkOperator.java:589)

at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:467)

at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:758)

at
org.apache.hadoop.hive.ql.exec.SelectOperator.processOp(SelectOperator.java:84)

at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:467)

at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:758)

at
org.apache.hadoop.hive.ql.exec.SelectOperator.processOp(SelectOperator.java:84)

at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:467)

at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:758)

at
org.apache.hadoop.hive.ql.exec.GroupByOperator.forward(GroupByOperator.java:964)

at
org.apache.hadoop.hive.ql.exec.GroupByOperator.processAggr(GroupByOperator.java:781)

at
org.apache.hadoop.hive.ql.exec.GroupByOperator.processOp(GroupByOperator.java:707)

... 5 more

Ended Job = job_local_0001 with errors

[2014-11-05 12:46:16,665] ERROR {org.apache.hadoop.hive.ql.exec.ExecDriver}
-  Ended Job = job_local_0001 with errors

Error during job, obtaining debugging information...

[2014-11-05 12:46:16,667] ERROR {org.apache.hadoop.hive.ql.exec.ExecDriver}
-  Error during job, obtaining debugging information...

Execution failed with exit status: 2

[2014-11-05 12:46:16,827] ERROR {org.apache.hadoop.hive.ql.exec.Task} -
Execution failed with exit status: 2

Obtaining error information

[2014-11-05 12:46:16,827] ERROR {org.apache.hadoop.hive.ql.exec.Task} -
Obtaining error information

Task failed!

Task ID:

  Stage-0

Logs:

[2014-11-05 12:46:16,827] ERROR {org.apache.hadoop.hive.ql.exec.Task} -

Task failed!

Task ID:

  Stage-0

Logs:

/Users/lakshman/Desktop/hiveBreaker/test2/wso2bam-2.5.0/repository/logs//wso2carbon.log

[2014-11-05 12:46:16,827] ERROR {org.apache.hadoop.hive.ql.exec.Task} -
/Users/lakshman/Desktop/hiveBreaker/test2/wso2bam-2.5.0/repository/logs//wso2carbon.log

[2014-11-05 12:46:16,827] ERROR {org.apache.hadoop.hive.ql.exec.ExecDriver}
-  Execution failed with exit status: 2

FAILED: Execution Error, return code 2 from
org.apache.hadoop.hive.ql.exec.MapRedTask

[2014-11-05 12:46:16,828] ERROR {org.apache.hadoop.hive.ql.Driver} -
FAILED: Execution Error, return code 2 from
org.apache.hadoop.hive.ql.exec.MapRedTask

[2014-11-05 12:46:16,832] ERROR
{org.wso2.carbon.analytics.hive.impl.HiveExecutorServiceImpl} -  Error
while executing Hive script.

Query returned non-zero code: 9, cause: FAILED: Execution Error, return
code 2 from org.apache.hadoop.hive.ql.exec.MapRedTask

java.sql.SQLException: Query returned non-zero code: 9, cause: FAILED:
Execution Error, return code 2 from
org.apache.hadoop.hive.ql.exec.MapRedTask

at
org.apache.hadoop.hive.jdbc.HiveStatement.executeQuery(HiveStatement.java:189)

at
org.wso2.carbon.analytics.hive.impl.HiveExecutorServiceImpl$ScriptCallable.executeHiveQuery(HiveExecutorServiceImpl.java:578)

at
org.wso2.carbon.analytics.hive.impl.HiveExecutorServiceImpl$ScriptCallable.call(HiveExecutorServiceImpl.java:286)

at
org.wso2.carbon.analytics.hive.impl.HiveExecutorServiceImpl$ScriptCallable.call(HiveExecutorServiceImpl.java:189)

at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303)

at java.util.concurrent.FutureTask.run(FutureTask.java:138)

at
java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:895)

at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:918)

at java.lang.Thread.run(Thread.java:695)

[2014-11-05 12:46:16,833] ERROR
{org.wso2.carbon.analytics.hive.task.HiveScriptExecutorTask} -  Error while
executing script : am_response_stats_analyzer

org.wso2.carbon.analytics.hive.exception.HiveExecutionException: Error
while executing Hive script.Query returned non-zero code: 9, cause: FAILED:
Execution Error, return code 2 from
org.apache.hadoop.hive.ql.exec.MapRedTask

at
org.wso2.carbon.analytics.hive.impl.HiveExecutorServiceImpl.execute(HiveExecutorServiceImpl.java:115)

at
org.wso2.carbon.analytics.hive.task.HiveScriptExecutorTask.execute(HiveScriptExecutorTask.java:70)

at
org.wso2.carbon.ntask.core.impl.TaskQuartzJobAdapter.execute(TaskQuartzJobAdapter.java:67)

at org.quartz.core.JobRunShell.run(JobRunShell.java:213)

at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:439)

at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303)

at java.util.concurrent.FutureTask.run(FutureTask.java:138)

at
java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:895)

at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:918)

at java.lang.Thread.run(Thread.java:695)

2014-11-05 12:46:17,620 null map = 100%,  reduce = 100%


[1] https://wso2.org/jira/browse/APIMANAGER-2992

[2]
https://drive.google.com/file/d/0B9KDy4GJKr1vNFp0NEVwbnZhbjg/view?usp=sharing
<https://drive.google.com/a/wso2.com/file/d/0Bx6Pq431GuEiSzBUVUxkNllnODQ/view?usp=sharing>

-- 
Lakshman Udayakantha
WSO2 Inc. www.wso2.com
lean.enterprise.middleware
Mobile: *0711241005*
_______________________________________________
Dev mailing list
[email protected]
http://wso2.org/cgi-bin/mailman/listinfo/dev

Reply via email to