Hi Sinthuja, Thanks for the quick response.
I did create the tables as per the instructions :-) But 'API_REQUEST_SUMMARY' table was gone after the auto-commit error, causing the subsequent error messages. I could reproduce it consistently. Thanks Rushhmin On Wed, Nov 4, 2015 at 2:09 PM, Sinthuja Ragendran <[email protected]> wrote: > Hi Rushmin, > > Spark JDBC connector first drops the existing table and recreates, but > anyhow to execute for the first time you need to have the table created. > Therefore please create the tables in the database by running mysql script > [1] which was given in the blog post. > > [1] > https://github.com/ruks/WSO2-APIM_DAS_Analytics_CApp/tree/v1.0.2/dbscripts > > Thanks, > Sinthuja. > > On Wed, Nov 4, 2015 at 7:24 PM, Rushmin Fernando <[email protected]> wrote: > >> Hi Devs, >> >> I was trying to publish APIM stats to DAS following this blog [1] >> >> And in DAS console I got the following error. >> >> [2015-11-04 12:10:08,163] ERROR >> {org.wso2.carbon.ntask.core.impl.TaskQuartzJobAdapter} - Error in >> executing task: Can't call commit when autocommit=true >> java.lang.RuntimeException: Can't call commit when autocommit=true >> at >> org.apache.spark.sql.jdbc.carbon.JDBCRelation.insert(JDBCRelation.scala:193) >> at >> org.apache.spark.sql.sources.InsertIntoDataSource.run(commands.scala:53) >> at >> org.apache.spark.sql.execution.ExecutedCommand.sideEffectResult$lzycompute(commands.scala:57) >> at >> org.apache.spark.sql.execution.ExecutedCommand.sideEffectResult(commands.scala:57) >> at >> org.apache.spark.sql.execution.ExecutedCommand.doExecute(commands.scala:68) >> at >> org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:88) >> at >> org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:88) >> at >> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:147) >> at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:87) >> at >> org.apache.spark.sql.SQLContext$QueryExecution.toRdd$lzycompute(SQLContext.scala:950) >> at >> org.apache.spark.sql.SQLContext$QueryExecution.toRdd(SQLContext.scala:950) >> at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:144) >> at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:128) >> at org.apache.spark.sql.DataFrame$.apply(DataFrame.scala:51) >> at org.apache.spark.sql.SQLContext.sql(SQLContext.scala:755) >> >> >> Immediately after that error the below error was there. >> >> [2015-11-04 12:15:00,027] ERROR >> {org.wso2.carbon.ntask.core.impl.TaskQuartzJobAdapter} - Error in >> executing task: Error while connecting to datasource WSO2AM_STATS_DB : >> Table 'TP_WSO2AM_STATS_DB.API_REQUEST_SUMMARY' doesn't exist >> java.lang.RuntimeException: Error while connecting to datasource >> WSO2AM_STATS_DB : Table 'TP_WSO2AM_STATS_DB.API_REQUEST_SUMMARY' doesn't >> exist >> at >> org.apache.spark.sql.jdbc.carbon.JDBCRelation.liftedTree1$1(JDBCRelation.scala:143) >> at >> org.apache.spark.sql.jdbc.carbon.JDBCRelation.<init>(JDBCRelation.scala:137) >> at >> org.apache.spark.sql.jdbc.carbon.AnalyticsJDBCRelationProvider.createRelation(JDBCRelation.scala:119) >> at org.apache.spark.sql.sources.ResolvedDataSource$.apply(ddl.scala:269) >> at org.apache.spark.sql.sources.CreateTempTableUsing.run(ddl.scala:412) >> at >> org.apache.spark.sql.execution.ExecutedCommand.sideEffectResult$lzycompute(commands.scala:57) >> at >> org.apache.spark.sql.execution.ExecutedCommand.sideEffectResult(commands.scala:57) >> at >> org.apache.spark.sql.execution.ExecutedCommand.doExecute(commands.scala:68) >> at >> org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:88) >> at >> org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:88) >> at >> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:147) >> at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:87) >> at >> org.apache.spark.sql.SQLContext$QueryExecution.toRdd$lzycompute(SQLContext.scala:950) >> at >> org.apache.spark.sql.SQLContext$QueryExecution.toRdd(SQLContext.scala:950) >> at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:144) >> at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:128) >> at org.apache.spark.sql.DataFrame$.apply(DataFrame.scala:51) >> at org.apache.spark.sql.SQLContext.sql(SQLContext.scala:755) >> >> >> As the error message says the 'TP_WSO2AM_STATS_DB.API_REQUEST_SUMMARY' >> table was gone ! >> >> *I added the relaxAutoCommit=true property to the JDBC connection string >> to solve the first error message and stat feature worked.* >> >> Is the table deletion, a result of the first error (trying to commit when >> auto commit is enabled) >> >> Thanks >> Rushmin >> >> >> >> >> [1] - >> http://blog.rukspot.com/2015/09/publishing-apim-runtime-statistics-to.html >> >> >> >> -- >> *Rushmin Fernando* >> *Technical Lead* >> >> WSO2 Inc. <http://wso2.com/> - Lean . Enterprise . Middleware >> >> email : [email protected] >> mobile : +94772310855 >> >> >> >> _______________________________________________ >> Dev mailing list >> [email protected] >> http://wso2.org/cgi-bin/mailman/listinfo/dev >> >> > > > -- > *Sinthuja Rajendran* > Associate Technical Lead > WSO2, Inc.:http://wso2.com > > Blog: http://sinthu-rajan.blogspot.com/ > Mobile: +94774273955 > > > -- *Rushmin Fernando* *Technical Lead* WSO2 Inc. <http://wso2.com/> - Lean . Enterprise . Middleware email : [email protected] mobile : +94772310855
_______________________________________________ Dev mailing list [email protected] http://wso2.org/cgi-bin/mailman/listinfo/dev
