Hi Marco, This error occur when the cartridge definition doesn't contain iaasProvider section. I guess the path to your cartridge json is incorrect in the curl command.
I deployed the same cartridge json, that you are using and I did not face any issue. Can you switch to JSON view from the UI and paste the cartridge definition and try to add it [1]. [1] https://drive.google.com/file/d/0B0v957zZwVWrQjVWX2x6YzI3Mkk/view?usp=sharing <https://drive.google.com/file/d/0B0v957zZwVWrQjVWX2x6YzI3Mkk/view?usp=sharing> On Wed, Sep 9, 2015 at 12:39 AM, Monaco Marco <[email protected]> wrote: > Anuruddha, > > thank you for the suggestion. > > I made a git pull from repo without deleting DBs (just replacing files) > and it started to spawn new instances, but it complained about public ip > addresses and stopping. > > i made a fresh installation deleting all SQL dbs, but now we are stucked > on another error. > > I successfully set up again Network Partition, Autoscale Policies, > Deployment Policies, but when I came to cartridge Stratos started to > missbehave. > > If I try to set up a cartridge on GUI it just stucks and then came back to > Add Cartridge page. > > I tryied to use api (https://127.0.0.1:9443/api/cartridges) sending this > json: > > { > "type": "wso2esb-481-manager", > "category": "framework", > "provider": "wso2", > "host": "esb.alma.it", > "displayName": "WSO2 ESB 4.8.1 Manager", > "description": "WSO2 ESB 4.8.1 Manager Cartridge", > "version": "4.8.1", > "multiTenant": false, > "loadBalancingIPType": "private", > "portMapping": [ > { > "name": "mgt-http", > "protocol": "http", > "port": 9763, > "proxyPort": 0 > }, > { > "name": "mgt-https", > "protocol": "https", > "port": 9443, > "proxyPort": 0 > }, > { > "name": "pt-http", > "protocol": "http", > "port": 8280, > "proxyPort": 0 > }, > { > "name": "pt-https", > "protocol": "https", > "port": 8243, > "proxyPort": 0 > } > ], > "iaasProvider": [ > { > "type": "openstack", > "imageId": "RegionOne/c2951a15-47b7-4f9c-a6e0-d3b7a50bc9aa", > "networkInterfaces": [ > { > "networkUuid": "bd02ca5c-4a57-45c3-8478-db0624829bdb" > } > ], > "property": [ > { > "name": "instanceType", > "value": "RegionOne/3" > }, > { > "name": "securityGroups", > "value": "default" > }, > { > "name": "autoAssignIp", > "value": "true" > }, > { > "name": "keyPair", > "value": "alma-keypair" > } > ] > } > ], > "property": [ > { > "name": "payload_parameter.CONFIG_PARAM_CLUSTERING", > "value": "true" > }, > { > "name": "payload_parameter.LB_IP", > "value": "<LOAD_BALANCER_IP>" > } > ] > } > > and getting this response: > > {"status":"error","message":"IaaS providers not found in cartridge: null"} > > in stratos logs I can see the same errors: > > [2015-09-08 19:00:26,600] ERROR > {org.apache.stratos.rest.endpoint.handlers.CustomExceptionMapper} - IaaS > providers not found in cartridge: null > org.apache.stratos.rest.endpoint.exception.RestAPIException: IaaS > providers not found in cartridge: null > at > org.apache.stratos.rest.endpoint.api.StratosApiV41Utils.addCartridge(StratosApiV41Utils.java:126) > at > org.apache.stratos.rest.endpoint.api.StratosApiV41.addCartridge(StratosApiV41.java:292) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:606) > at > org.apache.cxf.service.invoker.AbstractInvoker.performInvocation(AbstractInvoker.java:180) > at > org.apache.cxf.service.invoker.AbstractInvoker.invoke(AbstractInvoker.java:96) > at org.apache.cxf.jaxrs.JAXRSInvoker.invoke(JAXRSInvoker.java:194) > at org.apache.cxf.jaxrs.JAXRSInvoker.invoke(JAXRSInvoker.java:100) > at > org.apache.cxf.interceptor.ServiceInvokerInterceptor$1.run(ServiceInvokerInterceptor.java:57) > at > org.apache.cxf.interceptor.ServiceInvokerInterceptor.handleMessage(ServiceInvokerInterceptor.java:93) > at > org.apache.cxf.phase.PhaseInterceptorChain.doIntercept(PhaseInterceptorChain.java:271) > at > org.apache.cxf.transport.ChainInitiationObserver.onMessage(ChainInitiationObserver.java:121) > at > org.apache.cxf.transport.http.AbstractHTTPDestination.invoke(AbstractHTTPDestination.java:239) > at > org.apache.cxf.transport.servlet.ServletController.invokeDestination(ServletController.java:223) > at > org.apache.cxf.transport.servlet.ServletController.invoke(ServletController.java:203) > at > org.apache.cxf.transport.servlet.ServletController.invoke(ServletController.java:137) > at > org.apache.cxf.transport.servlet.CXFNonSpringServlet.invoke(CXFNonSpringServlet.java:159) > at > org.apache.cxf.transport.servlet.AbstractHTTPServlet.handleRequest(AbstractHTTPServlet.java:286) > at > org.apache.cxf.transport.servlet.AbstractHTTPServlet.doPost(AbstractHTTPServlet.java:206) > at javax.servlet.http.HttpServlet.service(HttpServlet.java:755) > at > org.apache.cxf.transport.servlet.AbstractHTTPServlet.service(AbstractHTTPServlet.java:262) > at > org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:303) > > > and i'm not able to add any cartridge, including simple php. > > I triple checked all IaaS settings in cloud-controller.xml and in > cartridge definitions and they are correct. > > Any throught? > > Tanks.... > > Marco > ------------------------------ > *Da:* Anuruddha Liyanarachchi [[email protected]] > *Inviato:* martedì 8 settembre 2015 15.46 > *A:* Monaco Marco; WSO2 Developers' List > *Cc:* [email protected] > *Oggetto:* Re: Strange Error in WSO2 Private Paas 4.1.2 > > [Removing stratos dev and adding wso2dev ] > > Hi Marco, > > We have fixed this in the master branch with commit [1]. Please take a > pull from master branch or download the alpha release pack from [2]. > > [1] > https://github.com/wso2/product-private-paas/commit/54a77e9d85538a20ace00c57ec9e8aed410fb773 > [2]https://svn.wso2.org/repos/wso2/scratch/PPAAS/wso2ppaas-4.1.0-ALPHA/ > > On Tue, Sep 8, 2015 at 7:06 PM, Monaco Marco <[email protected]> > wrote: > >> Hi, >> >> We have successfully installed WSO2 PPaaS 4.1.2 on our Openstack IaaS >> Environment. >> >> After following this procedure ( >> https://docs.wso2.com/display/PP410/Deploy+Private+PaaS+in+OpenStack) we >> are able to open the Private PaaS console and configure Network Partitions, >> Autoscale Policies, Deployment, Cartridge, ecc.. >> >> We have problems trying to deploy applications. We tested both PHP and >> WSO2ESB applications, also using Mock IaaS, but we receive always the same >> error: >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> *ERROR {org.apache.stratos.autoscaler.rule.AutoscalerRuleEvaluator} - >> Unable to Analyse Expression log.debug("[scaling] Number of required >> instances based on stats: " + numberOfRequiredInstances + " " + >> "[active instances count] " + activeInstancesCount + " >> [network-partition] " + >> clusterInstanceContext.getNetworkPartitionId() + " [cluster] " + >> clusterId); int nonTerminatedMembers = >> clusterInstanceContext.getNonTerminatedMemberCount(); if(scaleUp){ >> int clusterMaxMembers = >> clusterInstanceContext.getMaxInstanceCount(); if >> (nonTerminatedMembers < clusterMaxMembers) { int >> additionalInstances = 0; if(clusterMaxMembers < >> numberOfRequiredInstances){ additionalInstances = >> clusterMaxMembers - nonTerminatedMembers; log.info >> <http://log.info>("[scale-up] Required member count based on stat based >> scaling is higher than max, hence" + " >> notifying to parent for possible group scaling or app bursting. [cluster] " >> + clusterId + " [instance id]" + >> clusterInstanceContext.getId() + " [max] " + clusterMaxMembers >> + " [number of required instances] " + >> numberOfRequiredInstances + " [additional >> instances to be created] " + additionalInstances); >> delegator.delegateScalingOverMaxNotification(clusterId, >> clusterInstanceContext.getNetworkPartitionId(), >> clusterInstanceContext.getId()); } else { >> additionalInstances = numberOfRequiredInstances - >> nonTerminatedMembers; } >> clusterInstanceContext.resetScaleDownRequestsCount(); >> log.debug("[scale-up] " + " [has scaling dependents] " + >> clusterInstanceContext.hasScalingDependants() + " >> [cluster] " + clusterId ); >> if(clusterInstanceContext.hasScalingDependants()) { >> log.debug("[scale-up] Notifying dependencies [cluster] " + clusterId); >> >> delegator.delegateScalingDependencyNotification(clusterId, >> clusterInstanceContext.getNetworkPartitionId(), >> clusterInstanceContext.getId(), numberOfRequiredInstances, >> clusterInstanceContext.getMinInstanceCount()); } else { >> boolean partitionsAvailable = true; >> int count = 0; String autoscalingReason = >> (numberOfRequiredInstances == >> numberOfInstancesReuquiredBasedOnRif)?"Scaling up due to RIF, [Predicted >> Value] "+rifPredictedValue+" [Threshold] >> "+rifThreshold:(numberOfRequiredInstances== >> numberOfInstancesReuquiredBasedOnMemoryConsumption)?"Scaling up due to MC, >> [Predicted Value] "+mcPredictedValue+" [Threshold] "+mcThreshold:"Scaling >> up due to LA, [Predicted Value] "+laPredictedValue+" [Threshold] >> "+laThreshold; autoscalingReason += " [Number of >> required instances] "+numberOfRequiredInstances+" [Cluster Max Members] >> "+clusterMaxMembers+" [Additional instances to be created] " + >> additionalInstances; while(count != additionalInstances >> && partitionsAvailable){ >> ClusterLevelPartitionContext partitionContext = >> (ClusterLevelPartitionContext) >> partitionAlgorithm.getNextScaleUpPartitionContext(clusterInstanceContext.getPartitionCtxtsAsAnArray()); >> if(partitionContext != null){ >> log.info <http://log.info>("[scale-up] >> Partition available, hence trying to spawn an instance to scale up! " + >> " [application id] " + applicationId + >> " [cluster] " + clusterId + " [instance id] >> " + clusterInstanceContext.getId() + " >> [network-partition] " + clusterInstanceContext.getNetworkPartitionId() + >> " [partition] " + >> partitionContext.getPartitionId() + " >> scaleup due to RIF: " + (rifReset && (rifPredictedValue > rifThreshold)) + >> " [rifPredictedValue] " + rifPredictedValue >> + " [rifThreshold] " + rifThreshold + " >> scaleup due to MC: " + (mcReset && (mcPredictedValue > mcThreshold)) + >> " [mcPredictedValue] " + mcPredictedValue + >> " [mcThreshold] " + mcThreshold + " scaleup >> due to LA: " + (laReset && (laPredictedValue > laThreshold)) + >> " [laPredictedValue] " + laPredictedValue + >> " [laThreshold] " + laThreshold); >> log.debug("[scale-up] " + " [partition] " + >> partitionContext.getPartitionId() + " [cluster] " + clusterId ); >> long scalingTime = System.currentTimeMillis(); >> delegator.delegateSpawn(partitionContext, >> clusterId, clusterInstanceContext.getId(), >> isPrimary,autoscalingReason,scalingTime); >> count++; } else { >> log.warn("[scale-up] No more partition available even though " + >> "cartridge-max is not reached!, [cluster] " + >> clusterId + " Please update deployment-policy >> with new partitions or with higher " + >> "partition-max"); partitionsAvailable = false; >> } } } >> } else { log.info <http://log.info>("[scale-up] >> Trying to scale up over max, hence not scaling up cluster itself and >> notifying to parent for possible group scaling or >> app bursting. [cluster] " + clusterId + " [instance >> id]" + clusterInstanceContext.getId() + " [max] " + >> clusterMaxMembers); >> delegator.delegateScalingOverMaxNotification(clusterId, >> clusterInstanceContext.getNetworkPartitionId(), >> clusterInstanceContext.getId()); } } else >> if(scaleDown){ if(nonTerminatedMembers > >> clusterInstanceContext.getMinInstanceCount){ >> log.debug("[scale-down] Decided to Scale down [cluster] " + clusterId); >> if(clusterInstanceContext.getScaleDownRequestsCount() > 2 >> ){ log.debug("[scale-down] Reached scale down requests >> threshold [cluster] " + clusterId + " Count " + >> clusterInstanceContext.getScaleDownRequestsCount()); >> if(clusterInstanceContext.hasScalingDependants()) { >> log.debug("[scale-up] Notifying dependencies [cluster] " + clusterId); >> >> delegator.delegateScalingDependencyNotification(clusterId, >> clusterInstanceContext.getNetworkPartitionId(), >> clusterInstanceContext.getId(), numberOfRequiredInstances, >> clusterInstanceContext.getMinInstanceCount()); } else{ >> MemberStatsContext selectedMemberStatsContext = >> null; double lowestOverallLoad = 0.0; >> boolean foundAValue = false; >> ClusterLevelPartitionContext partitionContext = >> (ClusterLevelPartitionContext) >> partitionAlgorithm.getNextScaleDownPartitionContext(clusterInstanceContext.getPartitionCtxtsAsAnArray()); >> if(partitionContext != null){ >> log.info <http://log.info>("[scale-down] >> Partition available to scale down " + " >> [application id] " + applicationId + " >> [cluster] " + clusterId + " [instance id] " + >> clusterInstanceContext.getId() + " >> [network-partition] " + clusterInstanceContext.getNetworkPartitionId() + >> " [partition] " + >> partitionContext.getPartitionId() + " >> scaledown due to RIF: " + (rifReset && (rifPredictedValue < rifThreshold)) >> + " [rifPredictedValue] " + >> rifPredictedValue + " [rifThreshold] " + rifThreshold + >> " scaledown due to MC: " + (mcReset && >> (mcPredictedValue < mcThreshold)) + " >> [mcPredictedValue] " + mcPredictedValue + " [mcThreshold] " + mcThreshold + >> " scaledown due to LA: " + (laReset && >> (laPredictedValue < laThreshold)) + " >> [laPredictedValue] " + laPredictedValue + " [laThreshold] " + laThreshold >> ); // In partition >> context member stat context, all the primary members need to be >> // avoided being selected as the member to >> terminated for(MemberStatsContext >> memberStatsContext: partitionContext.getMemberStatsContexts().values()){ >> if( >> !primaryMembers.contains(memberStatsContext.getMemberId()) ) { >> LoadAverage loadAverage = >> memberStatsContext.getLoadAverage(); >> log.debug("[scale-down] " + " [cluster] " >> + clusterId + " [member] " + >> memberStatsContext.getMemberId() + " Load average: " + loadAverage); >> MemoryConsumption memoryConsumption = >> memberStatsContext.getMemoryConsumption(); >> log.debug("[scale-down] " + " [partition] " + >> partitionContext.getPartitionId() + " [cluster] " >> + clusterId + " [member] " + >> memberStatsContext.getMemberId() + " Memory consumption: " + >> memoryConsumption); >> double predictedCpu = >> delegator.getPredictedValueForNextMinute(loadAverage.getAverage(), >> >> loadAverage.getGradient(),loadAverage.getSecondDerivative(), 1); >> log.debug("[scale-down] " + " [partition] " >> + partitionContext.getPartitionId() + " [cluster] " >> + clusterId + " [member] " + >> memberStatsContext.getMemberId() + " Predicted CPU: " + predictedCpu); >> double predictedMemoryConsumption = >> delegator.getPredictedValueForNextMinute( >> >> memoryConsumption.getAverage(),memoryConsumption.getGradient(),memoryConsumption.getSecondDerivative(), >> 1); log.debug("[scale-down] " + " >> [partition] " + partitionContext.getPartitionId() + " [cluster] " >> + clusterId + " [member] " + >> memberStatsContext.getMemberId() + " Predicted memory consumption: " + >> predictedMemoryConsumption); >> double overallLoad = (predictedCpu + >> predictedMemoryConsumption) / 2; >> log.debug("[scale-down] " + " [partition] " + >> partitionContext.getPartitionId() + " [cluster] " >> + clusterId + " [member] " + >> memberStatsContext.getMemberId() + " Overall load: " + overallLoad); >> if(!foundAValue){ >> foundAValue = true; >> selectedMemberStatsContext = >> memberStatsContext; lowestOverallLoad = >> overallLoad; } else if(overallLoad < >> lowestOverallLoad){ >> selectedMemberStatsContext = memberStatsContext; >> lowestOverallLoad = overallLoad; >> } } >> } >> if(selectedMemberStatsContext != null) { >> log.info <http://log.info>("[scale-down] Trying to terminating an instace >> to scale down!" ); log.debug("[scale-down] >> " + " [partition] " + partitionContext.getPartitionId() + " [cluster] " >> + clusterId + " Member with lowest >> overall load: " + selectedMemberStatsContext.getMemberId()); >> >> delegator.delegateTerminate(partitionContext, >> selectedMemberStatsContext.getMemberId()); } >> } } } else{ >> log.debug("[scale-down] Not reached scale down >> requests threshold. " + clusterId + " Count " + >> clusterInstanceContext.getScaleDownRequestsCount()); >> clusterInstanceContext.increaseScaleDownRequestsCount(); } >> } else { log.debug("[scale-down] Min is >> reached, hence not scaling down [cluster] " + clusterId + " [instance id]" >> + clusterInstanceContext.getId()); >> //if(clusterInstanceContext.isInGroupScalingEnabledSubtree()){ >> >> delegator.delegateScalingDownBeyondMinNotification(clusterId, >> clusterInstanceContext.getNetworkPartitionId(), >> clusterInstanceContext.getId()); //} } >> } else{ log.debug("[scaling] No decision made to either scale >> up or scale down ... [cluster] " + clusterId + " [instance id]" >> + clusterInstanceContext.getId()); };: [Error: unable >> to resolve method using strict-mode: >> org.apache.stratos.autoscaler.rule.RuleTasksDelegator.delegateSpawn(org.apache.stratos.autoscaler.context.partition.ClusterLevelPartitionContext, >> java.lang.String, java.lang.String, java.lang.Boolean, java.lang.String, >> long)] [Near : {... delegator.delegateSpawn(partitionContext ....}] * >> Once we try to deploy the application, it hangs and it's impossible to >> remove it unless we erase and repopulate Stratos database. It comes up >> everytime we restart stratos (if we don't recreate DB), but in any case >> it's impossible to deploy any application. >> >> These are configurations that we used for WSO2 ESB (we configured puppet >> side correctly according to >> https://docs.wso2.com/display/PP410/Configuring+Puppet+Master and >> https://github.com/wso2/product-private-paas/tree/master/cartridges/templates-modules/wso2esb-4.8.1). >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> *Autoscaling policy: { "id": "Autoscaling-WSO2", >> "loadThresholds": { "requestsInFlight": { "threshold": >> 20 }, "memoryConsumption": { "threshold": 80 >> }, "loadAverage": { "threshold": 120 } >> } } DEPLOYMENT POLICY: { "id": "Deployment-WSO2", >> "networkPartitions": [ { "id": "NP1", >> "partitionAlgo": "round-robin", "partitions": [ >> { "id": "P1", "partitionMax": 5, >> "partitionMin": 1 } ] >> } ] } APPLICATION POLICY: { "id": "Application-WSO2", >> "algorithm": "one-after-another", "networkPartitions": [ "NP1" >> ], "properties": [ ] } CARTRIDGES: MANAGER: { "type": >> "wso2esb-481-manager", "category": "framework", "provider": "wso2", >> "host": "esb.alma.it <http://esb.alma.it>", "displayName": "WSO2 >> ESB 4.8.1 Manager", "description": "WSO2 ESB 4.8.1 Manager Cartridge", >> "version": "4.8.1", "multiTenant": false, >> "loadBalancingIPType": "public", "portMapping": [ { >> "name": "mgt-http", "protocol": "http", "port": >> 9763, "proxyPort": 0 }, { "name": >> "mgt-https", "protocol": "https", "port": 9443, >> "proxyPort": 0 }, { "name": >> "pt-http", "protocol": "http", "port": 8280, >> "proxyPort": 0 }, { "name": >> "pt-https", "protocol": "https", "port": 8243, >> "proxyPort": 0 } ], "iaasProvider": [ { >> "type": "openstack", "imageId": >> "RegionOne/c2951a15-47b7-4f9c-a6e0-d3b7a50bc9aa", "property": [ >> { "name": "instanceType", >> "value": "RegionOne/3" }, >> { "name": "keyPair", >> "value": "alma_admin_keypair" }, >> { "name": "securityGroups", >> "value": "default" } ], >> "networkInterfaces": [ { >> "networkUuid": "ea0edbc6-6d6d-4efe-b11c-7cb3cb78256f" } >> ] } ], "property": [ { >> "name": "payload_parameter.CONFIG_PARAM_CLUSTERING", "value": >> "true" }, { "name": "payload_parameter.LB_IP", >> "value": "<LOAD_BALANCER_IP>" } ] } WORKER: { >> "type": "wso2esb-481-worker", "category": "framework", "provider": >> "wso2", "host": "esb.alma.it <http://esb.alma.it>", "displayName": >> "WSO2 ESB 4.8.1 Worker", "description": "WSO2 ESB 4.8.1 Worker >> Cartridge", "version": "4.8.1", "multiTenant": false, >> "loadBalancingIPType": "public", "portMapping": [ { >> "name": "pt-http", "protocol": "http", "port": >> 8280, "proxyPort": 0 }, { "name": >> "pt-https", "protocol": "https", "port": 8243, >> "proxyPort": 0 } ], "iaasProvider": [ { >> "type": "openstack", "imageId": >> "RegionOne/c2951a15-47b7-4f9c-a6e0-d3b7a50bc9aa", "property": [ >> { "name": "instanceType", >> "value": "RegionOne/3" }, >> { "name": "keyPair", >> "value": "alma_admin_keypair" }, >> { "name": "securityGroups", >> "value": "default" } ], >> "networkInterfaces": [ { >> "networkUuid": "ea0edbc6-6d6d-4efe-b11c-7cb3cb78256f" } >> ] } ], "property": [ { >> "name": "payload_parameter.CONFIG_PARAM_CLUSTERING", "value": >> "true" }, { "name": "payload_parameter.LB_IP", >> "value": "<LOAD_BALANCER_IP>" } ] } GRORUPING: { >> "name": "wso2esb-481-group", "cartridges": [ >> "wso2esb-481-manager", "wso2esb-481-worker" ], >> "dependencies": { "startupOrders": [ { >> "aliases": [ "cartridge.wso2esb-481-manager", >> "cartridge.wso2esb-481-worker" ] >> } ] } } APPLICATION: { "applicationId": >> "wso2esb-481-application", "alias": "wso2esb-481-application", >> "multiTenant": true, "components": { "groups": [ { >> "name": "wso2esb-481-group", "alias": >> "wso2esb-481-group", "deploymentPolicy": "Deployment-WSO2", >> "groupMinInstances": 1, >> "groupMaxInstances": 1, "cartridges": [ >> { "type": "wso2esb-481-manager", >> "cartridgeMin": 1, >> "cartridgeMax": 1, "subscribableInfo": { >> "alias": "wso2esb-481-manager", >> "autoscalingPolicy": "Autoscaling-WSO2" >> } }, { >> "type": "wso2esb-481-worker", >> "cartridgeMin": 2, >> "cartridgeMax": 5, "subscribableInfo": { >> "alias": "wso2esb-481-worker", >> "autoscalingPolicy": "Autoscaling-WSO2" >> } } ] >> } ] } }* >> >> >> Can you please help us? we are stucked at this point. >> >> Thank you all, >> >> Marco >> > > > > -- > *Thanks and Regards,* > Anuruddha Lanka Liyanarachchi > Software Engineer - WSO2 > Mobile : +94 (0) 712762611 > Tel : +94 112 145 345 > a <[email protected]>[email protected] > -- *Thanks and Regards,* Anuruddha Lanka Liyanarachchi Software Engineer - WSO2 Mobile : +94 (0) 712762611 Tel : +94 112 145 345 a <[email protected]>[email protected]
_______________________________________________ Dev mailing list [email protected] http://wso2.org/cgi-bin/mailman/listinfo/dev
