I didn't realize there was a difference in services for the native HA config. Thanks for that information.
I did a fresh install and below is how I deployed the registry. [root@ose-ha-master-01 playbook-openshift]# oc status In project default on server https://ose-master.ha.os.example.com:8443 svc/kubernetes - 172.30.0.1 ports 443, 53, 53 View details with 'oc describe <resource>/<name>' or list everything with 'oc get all'. [root@ose-ha-master-01 playbook-openshift]# oc get SecurityContextConstraints privileged -o json | jq '.users' [ "system:serviceaccount:default:registry", "system:serviceaccount:default:router", "system:serviceaccount:openshift-infra:build-controller", "system:serviceaccount:management-infra:management-admin" ] [root@ose-ha-master-01 playbook-openshift]# oadm registry \ > --service-account=registry \ > --config=/etc/origin/master/admin.kubeconfig \ > --credentials=/etc/origin/master/openshift-registry.kubeconfig \ > --images='registry.access.redhat.com/openshift3/ose-${component}:${version}' > \ > --selector="region=infra" DeploymentConfig "docker-registry" created Service "docker-registry" created Other than creating the dc and the svc nothing seems to happen. [root@ose-ha-master-01 playbook-openshift]# oc get deploymentconfig --all-namespaces NAMESPACE NAME TRIGGERS LATEST default docker-registry ConfigChange 1 [root@ose-ha-master-01 playbook-openshift]# oc get pods --all-namespaces [root@ose-ha-master-01 playbook-openshift]# oc get svc NAME CLUSTER_IP EXTERNAL_IP PORT(S) SELECTOR AGE docker-registry 172.30.151.27 <none> 5000/TCP docker-registry=default 1m kubernetes 172.30.0.1 <none> 443/TCP,53/UDP,53/TCP <none> 16m There are no pods and this `ansible nodes -m command -a 'docker ps'` tells me there are no containers present anywhere. The nodes are schedulable. What else can I look for? [root@ose-ha-master-01 playbook-openshift]# oc get nodes NAME LABELS STATUS AGE ose-ha-master-01.example.com kubernetes.io/hostname=ose-ha-master-01.example.com,region=infra,zone=rhev Ready,SchedulingDisabled 28m ose-ha-master-02.example.com kubernetes.io/hostname=ose-ha-master-02.example.com,region=infra,zone=rhev Ready,SchedulingDisabled 28m ose-ha-master-03.example.com kubernetes.io/hostname=ose-ha-master-03.example.com,region=infra,zone=rhev Ready,SchedulingDisabled 28m ose-ha-node-01.example.com kubernetes.io/hostname=ose-ha-node-01.example.com,region=infra,zone=rhev Ready 28m ose-ha-node-02.example.com kubernetes.io/hostname=ose-ha-node-02.example.com,region=infra,zone=rhev Ready 28m ose-ha-node-03.example.com kubernetes.io/hostname=ose-ha-node-03.example.com,region=primary,zone=rhev Ready 28m ose-ha-node-04.example.com kubernetes.io/hostname=ose-ha-node-04.example.com,region=primary,zone=rhev Ready 28m ose-ha-node-05.example.com kubernetes.io/hostname=ose-ha-node-05.example.com,region=primary,zone=rhev Ready 28m ose-ha-node-06.example.com kubernetes.io/hostname=ose-ha-node-06.example.com,region=primary,zone=rhev Ready 28m ----- On Jan 31, 2016, at 1:50 PM, Andrew Butcher <[email protected]> wrote: > Hey Dale, > Two services are started when using the native ha method; > atomic-openshift-master-api and atomic-openshift-master-controllers. How did > you deploy the registry? Are there any failed pods? > On Sun, Jan 31, 2016 at 4:14 PM, Dale Bewley < [email protected] > wrote: >> I'm provisioning a Native HA OpenShift Enterprise 3.1 cluster using byo >> playbook. There are no failures, but at the end the master service is >> disabled. >> (Full hosts file here: https://gist.github.com/dlbewley/d7db07edb7fa6da72259 >> ) >> [masters] >> ose-ha-master-[01:03]. example.com >> [etcd] >> ose-ha-etcd-[01:03]. example.com >> [lb] >> ose-ha-lb-01.example.com >> [nodes] >> ose-ha-master-[01:03]. example.com openshift_node_labels="{'region': 'infra', >> 'zone': 'rhev'}" openshift_schedulable=False >> ose-ha-node-[01:02]. example.com openshift_node_labels="{'region': 'infra', >> 'zone': 'rhev'}" >> ose-ha-node-[03:06]. example.com openshift_node_labels="{'region': 'primary', >> 'zone': 'rhev'}" >> The playbook runs with no failures, `oc get nodes` and everything looks fine, >> and I can even login to the web console by way of the loadbalancer node with >> my >> LDAP credentials. Quite an impressive playbook. :) >> PLAY RECAP >> ******************************************************************** >> localhost : ok=18 changed=0 unreachable=0 failed=0 >> ose-ha-etcd-01.example.com : ok=191 changed=40 unreachable=0 failed=0 >> ose-ha-etcd-02.example.com : ok=89 changed=20 unreachable=0 failed=0 >> ose-ha-etcd-03.example.com : ok=89 changed=20 unreachable=0 failed=0 >> ose-ha-lb-01.example.com : ok=29 changed=7 unreachable=0 failed=0 >> ose-ha-master-01.example.com : ok=334 changed=73 unreachable=0 failed=0 >> ose-ha-master-02.example.com : ok=203 changed=47 unreachable=0 failed=0 >> ose-ha-master-03.example.com : ok=203 changed=47 unreachable=0 failed=0 >> ose-ha-node-01.example.com : ok=103 changed=24 unreachable=0 failed=0 >> ose-ha-node-02.example.com : ok=103 changed=24 unreachable=0 failed=0 >> ose-ha-node-03.example.com : ok=103 changed=24 unreachable=0 failed=0 >> ose-ha-node-04.example.com : ok=103 changed=24 unreachable=0 failed=0 >> ose-ha-node-05.example.com : ok=103 changed=24 unreachable=0 failed=0 >> ose-ha-node-06.example.com : ok=103 changed=24 unreachable=0 failed=0 >> However when I attempted to deploy the registry nothing happened. I then >> noticed >> that `atomic-openshift-master` service was never enabled by the playbook. >> It looks like >> https://github.com/openshift/openshift-ansible/blob/master/roles/openshift_master/tasks/main.yml#L274 >> is skipped because this conditional: `when: not openshift_master_ha | bool`. >> [root@ose-ha-master-01 ]# grep -A1 'Start and enable master' >> ansible-byo-2016-01-28.log >> TASK: [openshift_master | Start and enable master] >> **************************** >> skipping: [ ose-ha-master-01.example.com ] >> -- >> TASK: [openshift_master | Start and enable master api] >> ************************ >> changed: [ ose-ha-master-01.example.com ] >> -- >> TASK: [openshift_master | Start and enable master controller] >> ***************** >> changed: [ ose-ha-master-01.example.com ] >> -- >> TASK: [openshift_master | Start and enable master] >> **************************** >> skipping: [ ose-ha-master-02.example.com ] >> -- >> TASK: [openshift_master | Start and enable master api] >> ************************ >> changed: [ ose-ha-master-02.example.com ] >> -- >> TASK: [openshift_master | Start and enable master controller] >> ***************** >> changed: [ ose-ha-master-02.example.com ] >> -- >> TASK: [openshift_master | Start and enable master] >> **************************** >> skipping: [ ose-ha-master-03.example.com ] >> -- >> TASK: [openshift_master | Start and enable master api] >> ************************ >> changed: [ ose-ha-master-03.example.com ] >> -- >> TASK: [openshift_master | Start and enable master controller] >> ***************** >> changed: [ ose-ha-master-03.example.com ] >> Any idea why this is happening? As near as I can discern openshift_master_ha >> is >> true, because `"{{ groups.oo_masters_to_config | length > 1 }}"` and I see >> my 3 >> masters in that group. >> _______________________________________________ >> users mailing list >> [email protected] >> http://lists.openshift.redhat.com/openshiftmm/listinfo/users
_______________________________________________ users mailing list [email protected] http://lists.openshift.redhat.com/openshiftmm/listinfo/users
