Thank you Supun for that information. Regards, Malith
On Thu, Sep 22, 2016 at 11:41 AM, Supun Sethunga <sup...@wso2.com> wrote: > BTW, This issue has already been fixed in the analytics-apim master branch. > > Regards, > Supun > > On Wed, Sep 21, 2016 at 3:04 PM, Malith Munasinghe <mali...@wso2.com> > wrote: > >> Hi All, >> >> Thanks for the prompt responses we will do the needful. >> >> Regards, >> Malith >> >> On Wed, Sep 21, 2016 at 2:54 PM, Rukshan Premathunga <ruks...@wso2.com> >> wrote: >> >>> Hi Malith, >>> >>> cApp we provided to Analytics APIM will not work for the DAS because of >>> the changes happen to the DAS 3.1.0. Because of that we need to use >>> Analytics APIM or have to update capp with the above changes. >>> >>> Thanks and Regards. >>> >>> On Wed, Sep 21, 2016 at 2:48 PM, Niranda Perera <nira...@wso2.com> >>> wrote: >>> >>>> Hi Malith, >>>> >>>> Yes, correct! you need to change the script. >>>> >>>> Additionally, there are some changes in the carbonJdbc connector as >>>> well... so, you might need to watch out for it! >>>> >>>> Please check with the APIM team and ESB team whether we are doing a >>>> feature release with the DAS 310 changes? >>>> >>>> cheers >>>> >>>> On Wed, Sep 21, 2016 at 5:11 AM, Malith Munasinghe <mali...@wso2.com> >>>> wrote: >>>> >>>>> Hi All, >>>>> >>>>> While preparing a DAS 3.1.0 to run APIM Analytics I have added >>>>> features as in [1] >>>>> <https://docs.wso2.com/display/AM200/Installing+WSO2+APIM+Analytics+Features>. >>>>> After deploying the CApp for APIM Analytics I run in to below error. >>>>> According to the error that *incrementalProcessing *is not a valid >>>>> option. Also according to [2] >>>>> <https://docs.wso2.com/display/DAS310/Incremental+Processing> the >>>>> syntax to parse this option is *incrementalParams. *In order to get >>>>> DAS 3.1.0 to process APIM Analytics >>>>> do we have to change the scripts with this option as well ? >>>>> >>>>> >>>>> TID: [-1234] [] [2016-09-21 08:54:00,019] ERROR >>>>> {org.wso2.carbon.analytics.spark.core.CarbonAnalyticsProcessorService} >>>>> - Error while executing query : CREATE TEMPORARY TABLE >>>>> APIMGT_PERMINUTE_REQUEST_DATA USING CarbonAnalytics OPTIONS(tableName >>>>> "ORG_WSO2_APIMGT_STATISTICS_PERMINUTEREQUEST", schema " year INT >>>>> -i, month INT -i, day INT -i, hour INT -i, minute INT -i, consumerKey >>>>> STRING, context STRING, api_version STRING, api STRING, version STRING, >>>>> requestTime LONG, userId STRING, hostName STRING, apiPublisher STRING, >>>>> total_request_count LONG, resourceTemplate STRING, method STRING, >>>>> applicationName STRING, tenantDomain STRING, userAgent STRING, >>>>> resourcePath STRING, request INT, applicationId STRING, tier STRING, >>>>> throttledOut BOOLEAN, clientIp STRING, applicationOwner STRING, >>>>> _timestamp LONG -i", primaryKeys "year, month, day, hour, minute, >>>>> consumerKey, context, api_version, userId, hostName, apiPublisher, >>>>> resourceTemplate, method, userAgent, clientIp", incrementalProcessing >>>>> "APIMGT_PERMINUTE_REQUEST_DATA, HOUR", mergeSchema "false") >>>>> {org.wso2.carbon.analytics.spark.core.CarbonAnalyticsProcessorService} >>>>> org.wso2.carbon.analytics.spark.core.exception.AnalyticsExecutionException: >>>>> Exception in executing query CREATE TEMPORARY TABLE >>>>> APIMGT_PERMINUTE_REQUEST_DATA USING CarbonAnalytics OPTIONS(tableName >>>>> "ORG_WSO2_APIMGT_STATISTICS_PERMINUTEREQUEST", schema " year INT >>>>> -i, month INT -i, day INT -i, hour INT -i, minute INT -i, consumerKey >>>>> STRING, context STRING, api_version STRING, api STRING, version STRING, >>>>> requestTime LONG, userId STRING, hostName STRING, apiPublisher STRING, >>>>> total_request_count LONG, resourceTemplate STRING, method STRING, >>>>> applicationName STRING, tenantDomain STRING, userAgent STRING, >>>>> resourcePath STRING, request INT, applicationId STRING, tier STRING, >>>>> throttledOut BOOLEAN, clientIp STRING, applicationOwner STRING, >>>>> _timestamp LONG -i", primaryKeys "year, month, day, hour, minute, >>>>> consumerKey, context, api_version, userId, hostName, apiPublisher, >>>>> resourceTemplate, method, userAgent, clientIp", incrementalProcessing >>>>> "APIMGT_PERMINUTE_REQUEST_DATA, HOUR", mergeSchema "false") >>>>> at org.wso2.carbon.analytics.spark.core.internal.SparkAnalytics >>>>> Executor.executeQueryLocal(SparkAnalyticsExecutor.java:764) >>>>> at org.wso2.carbon.analytics.spark.core.internal.SparkAnalytics >>>>> Executor.executeQuery(SparkAnalyticsExecutor.java:721) >>>>> at org.wso2.carbon.analytics.spark.core.CarbonAnalyticsProcesso >>>>> rService.executeQuery(CarbonAnalyticsProcessorService.java:201) >>>>> at org.wso2.carbon.analytics.spark.core.CarbonAnalyticsProcesso >>>>> rService.executeScript(CarbonAnalyticsProcessorService.java:151) >>>>> at org.wso2.carbon.analytics.spark.core.AnalyticsTask.execute(A >>>>> nalyticsTask.java:60) >>>>> at org.wso2.carbon.ntask.core.impl.TaskQuartzJobAdapter.execute >>>>> (TaskQuartzJobAdapter.java:67) >>>>> at org.quartz.core.JobRunShell.run(JobRunShell.java:213) >>>>> at java.util.concurrent.Executors$RunnableAdapter.call(Executor >>>>> s.java:471) >>>>> at java.util.concurrent.FutureTask.run(FutureTask.java:262) >>>>> at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPool >>>>> Executor.java:1145) >>>>> at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoo >>>>> lExecutor.java:615) >>>>> at java.lang.Thread.run(Thread.java:745) >>>>> Caused by: java.lang.RuntimeException: Unknown options : >>>>> incrementalprocessing >>>>> at org.wso2.carbon.analytics.spark.core.sources.AnalyticsRelati >>>>> onProvider.checkParameters(AnalyticsRelationProvider.java:123) >>>>> at org.wso2.carbon.analytics.spark.core.sources.AnalyticsRelati >>>>> onProvider.setParameters(AnalyticsRelationProvider.java:113) >>>>> at org.wso2.carbon.analytics.spark.core.sources.AnalyticsRelati >>>>> onProvider.createRelation(AnalyticsRelationProvider.java:75) >>>>> at org.wso2.carbon.analytics.spark.core.sources.AnalyticsRelati >>>>> onProvider.createRelation(AnalyticsRelationProvider.java:45) >>>>> at org.apache.spark.sql.execution.datasources.ResolvedDataSourc >>>>> e$.apply(ResolvedDataSource.scala:158) >>>>> at org.apache.spark.sql.execution.datasources.CreateTempTableUs >>>>> ing.run(ddl.scala:92) >>>>> at org.apache.spark.sql.execution.ExecutedCommand.sideEffectRes >>>>> ult$lzycompute(commands.scala:58) >>>>> at org.apache.spark.sql.execution.ExecutedCommand.sideEffectRes >>>>> ult(commands.scala:56) >>>>> at org.apache.spark.sql.execution.ExecutedCommand.doExecute(com >>>>> mands.scala:70) >>>>> at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$5. >>>>> apply(SparkPlan.scala:132) >>>>> at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$5. >>>>> apply(SparkPlan.scala:130) >>>>> at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperati >>>>> onScope.scala:150) >>>>> at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.s >>>>> cala:130) >>>>> at org.apache.spark.sql.execution.QueryExecution.toRdd$lzycompu >>>>> te(QueryExecution.scala:55) >>>>> at org.apache.spark.sql.execution.QueryExecution.toRdd(QueryExe >>>>> cution.scala:55) >>>>> at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:145) >>>>> at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:130) >>>>> at org.apache.spark.sql.DataFrame$.apply(DataFrame.scala:52) >>>>> at org.apache.spark.sql.SQLContext.sql(SQLContext.scala:817) >>>>> at org.wso2.carbon.analytics.spark.core.internal.SparkAnalytics >>>>> Executor.executeQueryLocal(SparkAnalyticsExecutor.java:760) >>>>> ... 11 more >>>>> TID: [-1234] [] [2016-09-21 08:54:00,020] ERROR >>>>> {org.wso2.carbon.analytics.spark.core.AnalyticsTask} - Error while >>>>> executing the scheduled task for the script: >>>>> APIM_INCREMENTAL_PROCESSING_SCRIPT >>>>> {org.wso2.carbon.analytics.spark.core.AnalyticsTask} >>>>> org.wso2.carbon.analytics.spark.core.exception.AnalyticsExecutionException: >>>>> Exception in executing query CREATE TEMPORARY TABLE >>>>> APIMGT_PERMINUTE_REQUEST_DATA USING CarbonAnalytics OPTIONS(tableName >>>>> "ORG_WSO2_APIMGT_STATISTICS_PERMINUTEREQUEST", schema " year INT >>>>> -i, month INT -i, day INT -i, hour INT -i, minute INT -i, consumerKey >>>>> STRING, context STRING, api_version STRING, api STRING, version STRING, >>>>> requestTime LONG, userId STRING, hostName STRING, apiPublisher STRING, >>>>> total_request_count LONG, resourceTemplate STRING, method STRING, >>>>> applicationName STRING, tenantDomain STRING, userAgent STRING, >>>>> resourcePath STRING, request INT, applicationId STRING, tier STRING, >>>>> throttledOut BOOLEAN, clientIp STRING, applicationOwner STRING, >>>>> _timestamp LONG -i", primaryKeys "year, month, day, hour, minute, >>>>> consumerKey, context, api_version, userId, hostName, apiPublisher, >>>>> resourceTemplate, method, userAgent, clientIp", incrementalProcessing >>>>> "APIMGT_PERMINUTE_REQUEST_DATA, HOUR", mergeSchema "false") >>>>> at org.wso2.carbon.analytics.spark.core.internal.SparkAnalytics >>>>> Executor.executeQueryLocal(SparkAnalyticsExecutor.java:764) >>>>> at org.wso2.carbon.analytics.spark.core.internal.SparkAnalytics >>>>> Executor.executeQuery(SparkAnalyticsExecutor.java:721) >>>>> at org.wso2.carbon.analytics.spark.core.CarbonAnalyticsProcesso >>>>> rService.executeQuery(CarbonAnalyticsProcessorService.java:201) >>>>> at org.wso2.carbon.analytics.spark.core.CarbonAnalyticsProcesso >>>>> rService.executeScript(CarbonAnalyticsProcessorService.java:151) >>>>> at org.wso2.carbon.analytics.spark.core.AnalyticsTask.execute(A >>>>> nalyticsTask.java:60) >>>>> at org.wso2.carbon.ntask.core.impl.TaskQuartzJobAdapter.execute >>>>> (TaskQuartzJobAdapter.java:67) >>>>> at org.quartz.core.JobRunShell.run(JobRunShell.java:213) >>>>> at java.util.concurrent.Executors$RunnableAdapter.call(Executor >>>>> s.java:471) >>>>> at java.util.concurrent.FutureTask.run(FutureTask.java:262) >>>>> at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPool >>>>> Executor.java:1145) >>>>> at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoo >>>>> lExecutor.java:615) >>>>> at java.lang.Thread.run(Thread.java:745) >>>>> Caused by: java.lang.RuntimeException: Unknown options : >>>>> incrementalprocessing >>>>> at org.wso2.carbon.analytics.spark.core.sources.AnalyticsRelati >>>>> onProvider.checkParameters(AnalyticsRelationProvider.java:123) >>>>> at org.wso2.carbon.analytics.spark.core.sources.AnalyticsRelati >>>>> onProvider.setParameters(AnalyticsRelationProvider.java:113) >>>>> at org.wso2.carbon.analytics.spark.core.sources.AnalyticsRelati >>>>> onProvider.createRelation(AnalyticsRelationProvider.java:75) >>>>> at org.wso2.carbon.analytics.spark.core.sources.AnalyticsRelati >>>>> onProvider.createRelation(AnalyticsRelationProvider.java:45) >>>>> at org.apache.spark.sql.execution.datasources.ResolvedDataSourc >>>>> e$.apply(ResolvedDataSource.scala:158) >>>>> at org.apache.spark.sql.execution.datasources.CreateTempTableUs >>>>> ing.run(ddl.scala:92) >>>>> at org.apache.spark.sql.execution.ExecutedCommand.sideEffectRes >>>>> ult$lzycompute(commands.scala:58) >>>>> at org.apache.spark.sql.execution.ExecutedCommand.sideEffectRes >>>>> ult(commands.scala:56) >>>>> at org.apache.spark.sql.execution.ExecutedCommand.doExecute(com >>>>> mands.scala:70) >>>>> at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$5. >>>>> apply(SparkPlan.scala:132) >>>>> at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$5. >>>>> apply(SparkPlan.scala:130) >>>>> at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperati >>>>> onScope.scala:150) >>>>> at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.s >>>>> cala:130) >>>>> at org.apache.spark.sql.execution.QueryExecution.toRdd$lzycompu >>>>> te(QueryExecution.scala:55) >>>>> at org.apache.spark.sql.execution.QueryExecution.toRdd(QueryExe >>>>> cution.scala:55) >>>>> at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:145) >>>>> at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:130) >>>>> at org.apache.spark.sql.DataFrame$.apply(DataFrame.scala:52) >>>>> at org.apache.spark.sql.SQLContext.sql(SQLContext.scala:817) >>>>> at org.wso2.carbon.analytics.spark.core.internal.SparkAnalytics >>>>> Executor.executeQueryLocal(SparkAnalyticsExecutor.java:760) >>>>> ... 11 more >>>>> >>>>> >>>>> [1] https://docs.wso2.com/display/AM200/Installing+WSO2+APIM >>>>> +Analytics+Features >>>>> [2] https://docs.wso2.com/display/DAS310/Incremental+Processing >>>>> >>>>> Regards, >>>>> Malith >>>>> -- >>>>> Malith Munasinghe | Software Engineer >>>>> M: +94 (71) 9401122 >>>>> E: mali...@wso2.com >>>>> W: http://wso2.com >>>>> <http://wso2.com/signature> >>>>> >>>> >>>> >>>> >>>> -- >>>> *Niranda Perera* >>>> Software Engineer, WSO2 Inc. >>>> Mobile: +94-71-554-8430 >>>> Twitter: @n1r44 <https://twitter.com/N1R44> >>>> https://pythagoreanscript.wordpress.com/ >>>> >>> >>> >>> >>> -- >>> Rukshan Chathuranga. >>> Software Engineer. >>> WSO2, Inc. >>> >> >> >> >> -- >> Malith Munasinghe | Software Engineer >> M: +94 (71) 9401122 >> E: mali...@wso2.com >> W: http://wso2.com >> <http://wso2.com/signature> >> >> _______________________________________________ >> Dev mailing list >> Dev@wso2.org >> http://wso2.org/cgi-bin/mailman/listinfo/dev >> >> > > > -- > *Supun Sethunga* > Senior Software Engineer > WSO2, Inc. > http://wso2.com/ > lean | enterprise | middleware > Mobile : +94 716546324 > Blog: http://supunsetunga.blogspot.com > -- Malith Munasinghe | Software Engineer M: +94 (71) 9401122 E: mali...@wso2.com W: http://wso2.com <http://wso2.com/signature>
_______________________________________________ Dev mailing list Dev@wso2.org http://wso2.org/cgi-bin/mailman/listinfo/dev