@Rukshan, you may have to update the blog accordingly :-) (saying that
auto-commit should be disabled)

On Wed, Nov 4, 2015 at 3:23 PM, Rushmin Fernando <rush...@wso2.com> wrote:

> Hi Sinthuja,
>
> Yes the relaxAutoCommit parameter was not there in the beginning and the
> table drop was probably due to that.
>
> Having the relaxAutoCommit in place and recreating the tables solved the
> issue :-)
>
> Just wanted to know whether the auto-commit related error caused the table
> drop.
>
> Thank you for your clarification.
>
> Thanks
> Rushmin
>
> On Wed, Nov 4, 2015 at 2:20 PM, Sinthuja Ragendran <sinth...@wso2.com>
> wrote:
>
>> Hi Rushmin,
>>
>> Did you create the table after encountering the exception, and adding
>> the relaxAutoCommit parameter as well? Because during the exception phase
>> it could have got dropped, and now it's again searching for the non-
>> existing table. Hence I wanted to make sure, you ran the mysql script now,
>> and still seeing this exception.
>>
>> Thanks,
>> Sinthuja.
>>
>> On Wed, Nov 4, 2015 at 7:46 PM, Rushmin Fernando <rush...@wso2.com>
>> wrote:
>>
>>> Hi Sinthuja,
>>>
>>> Thanks for the quick response.
>>>
>>> I did create the tables as per the instructions :-)
>>>
>>> But 'API_REQUEST_SUMMARY' table was gone after the auto-commit error,
>>> causing the subsequent error messages.
>>>
>>> I could reproduce it consistently.
>>>
>>> Thanks
>>> Rushhmin
>>>
>>>
>>> On Wed, Nov 4, 2015 at 2:09 PM, Sinthuja Ragendran <sinth...@wso2.com>
>>> wrote:
>>>
>>>> Hi Rushmin,
>>>>
>>>> Spark JDBC connector first drops the existing table and recreates, but
>>>> anyhow to execute for the first time you need to have the table created.
>>>> Therefore please create the tables in the database by running mysql script
>>>> [1] which was given in the blog post.
>>>>
>>>> [1]
>>>> https://github.com/ruks/WSO2-APIM_DAS_Analytics_CApp/tree/v1.0.2/dbscripts
>>>>
>>>> Thanks,
>>>> Sinthuja.
>>>>
>>>> On Wed, Nov 4, 2015 at 7:24 PM, Rushmin Fernando <rush...@wso2.com>
>>>> wrote:
>>>>
>>>>> Hi Devs,
>>>>>
>>>>> I was trying to publish APIM stats to DAS following this blog [1]
>>>>>
>>>>> And in DAS console I got the following error.
>>>>>
>>>>> [2015-11-04 12:10:08,163] ERROR
>>>>> {org.wso2.carbon.ntask.core.impl.TaskQuartzJobAdapter} -  Error in
>>>>> executing task: Can't call commit when autocommit=true
>>>>> java.lang.RuntimeException: Can't call commit when autocommit=true
>>>>> at
>>>>> org.apache.spark.sql.jdbc.carbon.JDBCRelation.insert(JDBCRelation.scala:193)
>>>>> at
>>>>> org.apache.spark.sql.sources.InsertIntoDataSource.run(commands.scala:53)
>>>>> at
>>>>> org.apache.spark.sql.execution.ExecutedCommand.sideEffectResult$lzycompute(commands.scala:57)
>>>>> at
>>>>> org.apache.spark.sql.execution.ExecutedCommand.sideEffectResult(commands.scala:57)
>>>>> at
>>>>> org.apache.spark.sql.execution.ExecutedCommand.doExecute(commands.scala:68)
>>>>> at
>>>>> org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:88)
>>>>> at
>>>>> org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:88)
>>>>> at
>>>>> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:147)
>>>>> at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:87)
>>>>> at
>>>>> org.apache.spark.sql.SQLContext$QueryExecution.toRdd$lzycompute(SQLContext.scala:950)
>>>>> at
>>>>> org.apache.spark.sql.SQLContext$QueryExecution.toRdd(SQLContext.scala:950)
>>>>> at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:144)
>>>>> at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:128)
>>>>> at org.apache.spark.sql.DataFrame$.apply(DataFrame.scala:51)
>>>>> at org.apache.spark.sql.SQLContext.sql(SQLContext.scala:755)
>>>>>
>>>>>
>>>>> Immediately after that error the below error was there.
>>>>>
>>>>> [2015-11-04 12:15:00,027] ERROR
>>>>> {org.wso2.carbon.ntask.core.impl.TaskQuartzJobAdapter} -  Error in
>>>>> executing task: Error while connecting to datasource WSO2AM_STATS_DB :
>>>>> Table 'TP_WSO2AM_STATS_DB.API_REQUEST_SUMMARY' doesn't exist
>>>>> java.lang.RuntimeException: Error while connecting to datasource
>>>>> WSO2AM_STATS_DB : Table 'TP_WSO2AM_STATS_DB.API_REQUEST_SUMMARY' doesn't
>>>>> exist
>>>>> at
>>>>> org.apache.spark.sql.jdbc.carbon.JDBCRelation.liftedTree1$1(JDBCRelation.scala:143)
>>>>> at
>>>>> org.apache.spark.sql.jdbc.carbon.JDBCRelation.<init>(JDBCRelation.scala:137)
>>>>> at
>>>>> org.apache.spark.sql.jdbc.carbon.AnalyticsJDBCRelationProvider.createRelation(JDBCRelation.scala:119)
>>>>> at
>>>>> org.apache.spark.sql.sources.ResolvedDataSource$.apply(ddl.scala:269)
>>>>> at org.apache.spark.sql.sources.CreateTempTableUsing.run(ddl.scala:412)
>>>>> at
>>>>> org.apache.spark.sql.execution.ExecutedCommand.sideEffectResult$lzycompute(commands.scala:57)
>>>>> at
>>>>> org.apache.spark.sql.execution.ExecutedCommand.sideEffectResult(commands.scala:57)
>>>>> at
>>>>> org.apache.spark.sql.execution.ExecutedCommand.doExecute(commands.scala:68)
>>>>> at
>>>>> org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:88)
>>>>> at
>>>>> org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:88)
>>>>> at
>>>>> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:147)
>>>>> at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:87)
>>>>> at
>>>>> org.apache.spark.sql.SQLContext$QueryExecution.toRdd$lzycompute(SQLContext.scala:950)
>>>>> at
>>>>> org.apache.spark.sql.SQLContext$QueryExecution.toRdd(SQLContext.scala:950)
>>>>> at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:144)
>>>>> at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:128)
>>>>> at org.apache.spark.sql.DataFrame$.apply(DataFrame.scala:51)
>>>>> at org.apache.spark.sql.SQLContext.sql(SQLContext.scala:755)
>>>>>
>>>>>
>>>>> As the error message says the 'TP_WSO2AM_STATS_DB.API_REQUEST_SUMMARY'
>>>>> table was gone !
>>>>>
>>>>> *I added the relaxAutoCommit=true property to the JDBC connection
>>>>> string to solve the first error message and stat feature worked.*
>>>>>
>>>>> Is the table deletion, a result of the first error (trying to commit
>>>>> when auto commit is enabled)
>>>>>
>>>>> Thanks
>>>>> Rushmin
>>>>>
>>>>>
>>>>>
>>>>>
>>>>> [1] -
>>>>> http://blog.rukspot.com/2015/09/publishing-apim-runtime-statistics-to.html
>>>>>
>>>>>
>>>>>
>>>>> --
>>>>> *Rushmin Fernando*
>>>>> *Technical Lead*
>>>>>
>>>>> WSO2 Inc. <http://wso2.com/> - Lean . Enterprise . Middleware
>>>>>
>>>>> email : rush...@wso2.com
>>>>> mobile : +94772310855
>>>>>
>>>>>
>>>>>
>>>>> _______________________________________________
>>>>> Dev mailing list
>>>>> Dev@wso2.org
>>>>> http://wso2.org/cgi-bin/mailman/listinfo/dev
>>>>>
>>>>>
>>>>
>>>>
>>>> --
>>>> *Sinthuja Rajendran*
>>>> Associate Technical Lead
>>>> WSO2, Inc.:http://wso2.com
>>>>
>>>> Blog: http://sinthu-rajan.blogspot.com/
>>>> Mobile: +94774273955
>>>>
>>>>
>>>>
>>>
>>>
>>> --
>>> *Rushmin Fernando*
>>> *Technical Lead*
>>>
>>> WSO2 Inc. <http://wso2.com/> - Lean . Enterprise . Middleware
>>>
>>> email : rush...@wso2.com
>>> mobile : +94772310855
>>>
>>>
>>>
>>
>>
>> --
>> *Sinthuja Rajendran*
>> Associate Technical Lead
>> WSO2, Inc.:http://wso2.com
>>
>> Blog: http://sinthu-rajan.blogspot.com/
>> Mobile: +94774273955
>>
>>
>>
>
>
> --
> *Rushmin Fernando*
> *Technical Lead*
>
> WSO2 Inc. <http://wso2.com/> - Lean . Enterprise . Middleware
>
> email : rush...@wso2.com
> mobile : +94772310855
>
>
>


-- 
*Rushmin Fernando*
*Technical Lead*

WSO2 Inc. <http://wso2.com/> - Lean . Enterprise . Middleware

email : rush...@wso2.com
mobile : +94772310855
_______________________________________________
Dev mailing list
Dev@wso2.org
http://wso2.org/cgi-bin/mailman/listinfo/dev

Reply via email to