I noted one other thing in the log:
TID: [-1234] [] [2016-02-02 12:43:53,909] INFO
{org.wso2.carbon.membership.scheme.kubernetes.KubernetesMembershipScheme} -
Kubernetes clustering configuration: [master] http://172.17.8.101:8080
[namespace] default [services] wso2as-manager
{org.wso2.carbon.membership.scheme.kubernetes.KubernetesMembershipScheme}
K8S services only has manager service. See the highlighted text. We need to
add both manager and worker here to let them connect to the same carbon
cluster.
Thanks
On Wed, Feb 3, 2016 at 12:19 AM, Vishanth Balasubramaniam <
[email protected]> wrote:
> Hi,
>
> I have been working on WSO2 AS worker/manager separated cluster in
> Kubernetes. I have setup the Kubernetes cluster using the Vagrant and
> CoreOs [1].
> I built the docker images for manager and worker with the following
> Kubernetes membership scheme configurations in the profile.
>
> wso2::clustering :
>> enabled : true
>> local_member_host : local.as.wso2.com
>> local_member_port : 4000
>> membership_scheme : kubernetes
>> k8 :
>> k8_master : http://172.17.8.101:8080
>> k8_namespace : default
>> k8_services : wso2as-manager,wso2as-worker
>> subDomain : mgt
>
>
> wso2::clustering :
>> enabled : true
>> local_member_host : worker.as.wso2.com
>> local_member_port : 4000
>> membership_scheme : kubernetes
>> k8 :
>> k8_master : http://172.17.8.101:8080
>> k8_namespace : default
>> k8_services : wso2as-manager,wso2as-worker
>> subDomain : worker
>
>
> I SCP-ed the saved zip file of built images and loaded them in the
> Kubernetes minion. Then I deployed the worker/manager service and
> controller. I SSH-ed into the container where the manager node is running,
> and there are no error logs as you can see below.
>
> TID: [-1234] [] [2016-02-02 12:43:53,892] INFO
>> {org.wso2.carbon.core.clustering.hazelcast.HazelcastClusteringAgent} -
>> Using kubernetes based membership management scheme
>> {org.wso2.carbon.core.clustering.hazelcast.HazelcastClusteringAgent}
>> TID: [-1234] [] [2016-02-02 12:43:53,905] INFO
>> {org.wso2.carbon.membership.scheme.kubernetes.KubernetesMembershipScheme}
>> - Initializing kubernetes membership scheme...
>> {org.wso2.carbon.membership.scheme.kubernetes.KubernetesMembershipScheme}
>> TID: [-1234] [] [2016-02-02 12:43:53,909] INFO
>> {org.wso2.carbon.membership.scheme.kubernetes.KubernetesMembershipScheme}
>> - Kubernetes clustering configuration: [master] http://172.17.8.101:8080
>> [namespace] default [services] wso2as-manager
>> {org.wso2.carbon.membership.scheme.kubernetes.KubernetesMembershipScheme}
>> TID: [-1234] [] [2016-02-02 12:43:54,499] INFO
>> {org.wso2.carbon.ui.internal.CarbonUIServiceComponent} - Mgt Console URL
>> : https://10.244.78.4:9443/carbon/
>> {org.wso2.carbon.ui.internal.CarbonUIServiceComponent}
>
>
> But it doesn't become the cluster coordinator node and there is no any
> other log after this. Also I am not able to access the carbon management
> console from my local machine.
>
> But whereas, when I deploy only the default profile (not cluster), then it
> runs fine and I am able to access the carbon management console from my
> local machine with the provided nodeport.
>
> Have I misconfigured anything in worker/manager cluster setup?
>
> [1] - https://github.com/pires/kubernetes-vagrant-coreos-cluster
>
> Regards,
> Vishanth
>
> --
> *Vishanth Balasubramaniam*
> Committer & PMC Member, Apache Stratos,
> Software Engineer, WSO2 Inc.; http://wso2.com
>
> mobile: *+94 77 17 377 18*
> about me: *http://about.me/vishanth <http://about.me/vishanth>*
>
--
*Imesh Gunaratne*
Senior Technical Lead
WSO2 Inc: http://wso2.com
T: +94 11 214 5345 M: +94 77 374 2057
W: http://imesh.gunaratne.org
Lean . Enterprise . Middleware
_______________________________________________
Dev mailing list
[email protected]
http://wso2.org/cgi-bin/mailman/listinfo/dev