BTW, This issue has already been fixed in the analytics-apim master branch.

Regards,
Supun

On Wed, Sep 21, 2016 at 3:04 PM, Malith Munasinghe <mali...@wso2.com> wrote:

> Hi All,
>
> Thanks for the prompt responses we will do the needful.
>
> Regards,
> Malith
>
> On Wed, Sep 21, 2016 at 2:54 PM, Rukshan Premathunga <ruks...@wso2.com>
> wrote:
>
>> Hi Malith,
>>
>> cApp we provided to Analytics APIM will not work for the DAS because of
>> the changes happen to the DAS 3.1.0. Because of that we need to use
>> Analytics APIM or have to update capp with the above changes.
>>
>> Thanks and Regards.
>>
>> On Wed, Sep 21, 2016 at 2:48 PM, Niranda Perera <nira...@wso2.com> wrote:
>>
>>> Hi Malith,
>>>
>>> Yes, correct! you need to change the script.
>>>
>>> Additionally, there are some changes in the carbonJdbc connector as
>>> well... so, you might need to watch out for it!
>>>
>>> Please check with the APIM team and ESB team whether we are doing a
>>> feature release with the DAS 310 changes?
>>>
>>> cheers
>>>
>>> On Wed, Sep 21, 2016 at 5:11 AM, Malith Munasinghe <mali...@wso2.com>
>>> wrote:
>>>
>>>> Hi All,
>>>>
>>>> While preparing a DAS 3.1.0 to run APIM Analytics I have added features
>>>> as in [1]
>>>> <https://docs.wso2.com/display/AM200/Installing+WSO2+APIM+Analytics+Features>.
>>>> After deploying the CApp for APIM Analytics I run in to below error.
>>>> According to the error that *incrementalProcessing *is not a valid
>>>> option. Also according to [2]
>>>> <https://docs.wso2.com/display/DAS310/Incremental+Processing> the
>>>> syntax to parse this option is *incrementalParams. *In order to get
>>>> DAS 3.1.0 to process APIM Analytics
>>>> do we have to change the scripts with this option as well ?
>>>>
>>>>
>>>> TID: [-1234] [] [2016-09-21 08:54:00,019] ERROR
>>>> {org.wso2.carbon.analytics.spark.core.CarbonAnalyticsProcessorService}
>>>> -  Error while executing query :         CREATE TEMPORARY TABLE
>>>> APIMGT_PERMINUTE_REQUEST_DATA USING CarbonAnalytics OPTIONS(tableName
>>>> "ORG_WSO2_APIMGT_STATISTICS_PERMINUTEREQUEST", schema "    year INT
>>>> -i, month INT -i, day INT -i, hour INT -i, minute INT -i,    consumerKey
>>>> STRING, context STRING, api_version STRING, api STRING, version STRING,
>>>> requestTime LONG, userId STRING, hostName STRING,    apiPublisher STRING,
>>>> total_request_count LONG, resourceTemplate STRING, method STRING,
>>>> applicationName STRING, tenantDomain STRING,    userAgent STRING,
>>>> resourcePath STRING, request INT, applicationId STRING, tier STRING,
>>>> throttledOut BOOLEAN, clientIp STRING,    applicationOwner STRING,
>>>> _timestamp LONG -i",    primaryKeys "year, month, day, hour, minute,
>>>> consumerKey, context, api_version, userId, hostName, apiPublisher,
>>>> resourceTemplate, method, userAgent, clientIp",    incrementalProcessing
>>>> "APIMGT_PERMINUTE_REQUEST_DATA, HOUR",    mergeSchema "false")
>>>> {org.wso2.carbon.analytics.spark.core.CarbonAnalyticsProcessorService}
>>>> org.wso2.carbon.analytics.spark.core.exception.AnalyticsExecutionException:
>>>> Exception in executing query CREATE TEMPORARY TABLE
>>>> APIMGT_PERMINUTE_REQUEST_DATA USING CarbonAnalytics OPTIONS(tableName
>>>> "ORG_WSO2_APIMGT_STATISTICS_PERMINUTEREQUEST", schema "    year INT
>>>> -i, month INT -i, day INT -i, hour INT -i, minute INT -i,    consumerKey
>>>> STRING, context STRING, api_version STRING, api STRING, version STRING,
>>>> requestTime LONG, userId STRING, hostName STRING,    apiPublisher STRING,
>>>> total_request_count LONG, resourceTemplate STRING, method STRING,
>>>> applicationName STRING, tenantDomain STRING,    userAgent STRING,
>>>> resourcePath STRING, request INT, applicationId STRING, tier STRING,
>>>> throttledOut BOOLEAN, clientIp STRING,    applicationOwner STRING,
>>>> _timestamp LONG -i",    primaryKeys "year, month, day, hour, minute,
>>>> consumerKey, context, api_version, userId, hostName, apiPublisher,
>>>> resourceTemplate, method, userAgent, clientIp",    incrementalProcessing
>>>> "APIMGT_PERMINUTE_REQUEST_DATA, HOUR",    mergeSchema "false")
>>>> at org.wso2.carbon.analytics.spark.core.internal.SparkAnalytics
>>>> Executor.executeQueryLocal(SparkAnalyticsExecutor.java:764)
>>>> at org.wso2.carbon.analytics.spark.core.internal.SparkAnalytics
>>>> Executor.executeQuery(SparkAnalyticsExecutor.java:721)
>>>> at org.wso2.carbon.analytics.spark.core.CarbonAnalyticsProcesso
>>>> rService.executeQuery(CarbonAnalyticsProcessorService.java:201)
>>>> at org.wso2.carbon.analytics.spark.core.CarbonAnalyticsProcesso
>>>> rService.executeScript(CarbonAnalyticsProcessorService.java:151)
>>>> at org.wso2.carbon.analytics.spark.core.AnalyticsTask.execute(A
>>>> nalyticsTask.java:60)
>>>> at org.wso2.carbon.ntask.core.impl.TaskQuartzJobAdapter.execute
>>>> (TaskQuartzJobAdapter.java:67)
>>>> at org.quartz.core.JobRunShell.run(JobRunShell.java:213)
>>>> at java.util.concurrent.Executors$RunnableAdapter.call(Executor
>>>> s.java:471)
>>>> at java.util.concurrent.FutureTask.run(FutureTask.java:262)
>>>> at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPool
>>>> Executor.java:1145)
>>>> at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoo
>>>> lExecutor.java:615)
>>>> at java.lang.Thread.run(Thread.java:745)
>>>> Caused by: java.lang.RuntimeException: Unknown options :
>>>> incrementalprocessing
>>>> at org.wso2.carbon.analytics.spark.core.sources.AnalyticsRelati
>>>> onProvider.checkParameters(AnalyticsRelationProvider.java:123)
>>>> at org.wso2.carbon.analytics.spark.core.sources.AnalyticsRelati
>>>> onProvider.setParameters(AnalyticsRelationProvider.java:113)
>>>> at org.wso2.carbon.analytics.spark.core.sources.AnalyticsRelati
>>>> onProvider.createRelation(AnalyticsRelationProvider.java:75)
>>>> at org.wso2.carbon.analytics.spark.core.sources.AnalyticsRelati
>>>> onProvider.createRelation(AnalyticsRelationProvider.java:45)
>>>> at org.apache.spark.sql.execution.datasources.ResolvedDataSourc
>>>> e$.apply(ResolvedDataSource.scala:158)
>>>> at org.apache.spark.sql.execution.datasources.CreateTempTableUs
>>>> ing.run(ddl.scala:92)
>>>> at org.apache.spark.sql.execution.ExecutedCommand.sideEffectRes
>>>> ult$lzycompute(commands.scala:58)
>>>> at org.apache.spark.sql.execution.ExecutedCommand.sideEffectRes
>>>> ult(commands.scala:56)
>>>> at org.apache.spark.sql.execution.ExecutedCommand.doExecute(com
>>>> mands.scala:70)
>>>> at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$5.
>>>> apply(SparkPlan.scala:132)
>>>> at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$5.
>>>> apply(SparkPlan.scala:130)
>>>> at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperati
>>>> onScope.scala:150)
>>>> at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.s
>>>> cala:130)
>>>> at org.apache.spark.sql.execution.QueryExecution.toRdd$lzycompu
>>>> te(QueryExecution.scala:55)
>>>> at org.apache.spark.sql.execution.QueryExecution.toRdd(QueryExe
>>>> cution.scala:55)
>>>> at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:145)
>>>> at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:130)
>>>> at org.apache.spark.sql.DataFrame$.apply(DataFrame.scala:52)
>>>> at org.apache.spark.sql.SQLContext.sql(SQLContext.scala:817)
>>>> at org.wso2.carbon.analytics.spark.core.internal.SparkAnalytics
>>>> Executor.executeQueryLocal(SparkAnalyticsExecutor.java:760)
>>>> ... 11 more
>>>> TID: [-1234] [] [2016-09-21 08:54:00,020] ERROR
>>>> {org.wso2.carbon.analytics.spark.core.AnalyticsTask} -  Error while
>>>> executing the scheduled task for the script: 
>>>> APIM_INCREMENTAL_PROCESSING_SCRIPT
>>>> {org.wso2.carbon.analytics.spark.core.AnalyticsTask}
>>>> org.wso2.carbon.analytics.spark.core.exception.AnalyticsExecutionException:
>>>> Exception in executing query CREATE TEMPORARY TABLE
>>>> APIMGT_PERMINUTE_REQUEST_DATA USING CarbonAnalytics OPTIONS(tableName
>>>> "ORG_WSO2_APIMGT_STATISTICS_PERMINUTEREQUEST", schema "    year INT
>>>> -i, month INT -i, day INT -i, hour INT -i, minute INT -i,    consumerKey
>>>> STRING, context STRING, api_version STRING, api STRING, version STRING,
>>>> requestTime LONG, userId STRING, hostName STRING,    apiPublisher STRING,
>>>> total_request_count LONG, resourceTemplate STRING, method STRING,
>>>> applicationName STRING, tenantDomain STRING,    userAgent STRING,
>>>> resourcePath STRING, request INT, applicationId STRING, tier STRING,
>>>> throttledOut BOOLEAN, clientIp STRING,    applicationOwner STRING,
>>>> _timestamp LONG -i",    primaryKeys "year, month, day, hour, minute,
>>>> consumerKey, context, api_version, userId, hostName, apiPublisher,
>>>> resourceTemplate, method, userAgent, clientIp",    incrementalProcessing
>>>> "APIMGT_PERMINUTE_REQUEST_DATA, HOUR",    mergeSchema "false")
>>>> at org.wso2.carbon.analytics.spark.core.internal.SparkAnalytics
>>>> Executor.executeQueryLocal(SparkAnalyticsExecutor.java:764)
>>>> at org.wso2.carbon.analytics.spark.core.internal.SparkAnalytics
>>>> Executor.executeQuery(SparkAnalyticsExecutor.java:721)
>>>> at org.wso2.carbon.analytics.spark.core.CarbonAnalyticsProcesso
>>>> rService.executeQuery(CarbonAnalyticsProcessorService.java:201)
>>>> at org.wso2.carbon.analytics.spark.core.CarbonAnalyticsProcesso
>>>> rService.executeScript(CarbonAnalyticsProcessorService.java:151)
>>>> at org.wso2.carbon.analytics.spark.core.AnalyticsTask.execute(A
>>>> nalyticsTask.java:60)
>>>> at org.wso2.carbon.ntask.core.impl.TaskQuartzJobAdapter.execute
>>>> (TaskQuartzJobAdapter.java:67)
>>>> at org.quartz.core.JobRunShell.run(JobRunShell.java:213)
>>>> at java.util.concurrent.Executors$RunnableAdapter.call(Executor
>>>> s.java:471)
>>>> at java.util.concurrent.FutureTask.run(FutureTask.java:262)
>>>> at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPool
>>>> Executor.java:1145)
>>>> at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoo
>>>> lExecutor.java:615)
>>>> at java.lang.Thread.run(Thread.java:745)
>>>> Caused by: java.lang.RuntimeException: Unknown options :
>>>> incrementalprocessing
>>>> at org.wso2.carbon.analytics.spark.core.sources.AnalyticsRelati
>>>> onProvider.checkParameters(AnalyticsRelationProvider.java:123)
>>>> at org.wso2.carbon.analytics.spark.core.sources.AnalyticsRelati
>>>> onProvider.setParameters(AnalyticsRelationProvider.java:113)
>>>> at org.wso2.carbon.analytics.spark.core.sources.AnalyticsRelati
>>>> onProvider.createRelation(AnalyticsRelationProvider.java:75)
>>>> at org.wso2.carbon.analytics.spark.core.sources.AnalyticsRelati
>>>> onProvider.createRelation(AnalyticsRelationProvider.java:45)
>>>> at org.apache.spark.sql.execution.datasources.ResolvedDataSourc
>>>> e$.apply(ResolvedDataSource.scala:158)
>>>> at org.apache.spark.sql.execution.datasources.CreateTempTableUs
>>>> ing.run(ddl.scala:92)
>>>> at org.apache.spark.sql.execution.ExecutedCommand.sideEffectRes
>>>> ult$lzycompute(commands.scala:58)
>>>> at org.apache.spark.sql.execution.ExecutedCommand.sideEffectRes
>>>> ult(commands.scala:56)
>>>> at org.apache.spark.sql.execution.ExecutedCommand.doExecute(com
>>>> mands.scala:70)
>>>> at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$5.
>>>> apply(SparkPlan.scala:132)
>>>> at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$5.
>>>> apply(SparkPlan.scala:130)
>>>> at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperati
>>>> onScope.scala:150)
>>>> at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.s
>>>> cala:130)
>>>> at org.apache.spark.sql.execution.QueryExecution.toRdd$lzycompu
>>>> te(QueryExecution.scala:55)
>>>> at org.apache.spark.sql.execution.QueryExecution.toRdd(QueryExe
>>>> cution.scala:55)
>>>> at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:145)
>>>> at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:130)
>>>> at org.apache.spark.sql.DataFrame$.apply(DataFrame.scala:52)
>>>> at org.apache.spark.sql.SQLContext.sql(SQLContext.scala:817)
>>>> at org.wso2.carbon.analytics.spark.core.internal.SparkAnalytics
>>>> Executor.executeQueryLocal(SparkAnalyticsExecutor.java:760)
>>>> ... 11 more
>>>>
>>>>
>>>> [1] https://docs.wso2.com/display/AM200/Installing+WSO2+APIM
>>>> +Analytics+Features
>>>> [2] https://docs.wso2.com/display/DAS310/Incremental+Processing
>>>>
>>>> Regards,
>>>> Malith
>>>> --
>>>> Malith Munasinghe | Software Engineer
>>>> M: +94 (71) 9401122
>>>> E: mali...@wso2.com
>>>> W: http://wso2.com
>>>> <http://wso2.com/signature>
>>>>
>>>
>>>
>>>
>>> --
>>> *Niranda Perera*
>>> Software Engineer, WSO2 Inc.
>>> Mobile: +94-71-554-8430
>>> Twitter: @n1r44 <https://twitter.com/N1R44>
>>> https://pythagoreanscript.wordpress.com/
>>>
>>
>>
>>
>> --
>> Rukshan Chathuranga.
>> Software Engineer.
>> WSO2, Inc.
>>
>
>
>
> --
> Malith Munasinghe | Software Engineer
> M: +94 (71) 9401122
> E: mali...@wso2.com
> W: http://wso2.com
> <http://wso2.com/signature>
>
> _______________________________________________
> Dev mailing list
> Dev@wso2.org
> http://wso2.org/cgi-bin/mailman/listinfo/dev
>
>


-- 
*Supun Sethunga*
Senior Software Engineer
WSO2, Inc.
http://wso2.com/
lean | enterprise | middleware
Mobile : +94 716546324
Blog: http://supunsetunga.blogspot.com
_______________________________________________
Dev mailing list
Dev@wso2.org
http://wso2.org/cgi-bin/mailman/listinfo/dev

Reply via email to