I didn't realize there was a difference in services for the native HA config. 
Thanks for that information. 

I did a fresh install and below is how I deployed the registry. 

[root@ose-ha-master-01 playbook-openshift]# oc status 
In project default on server https://ose-master.ha.os.example.com:8443 

svc/kubernetes - 172.30.0.1 ports 443, 53, 53 

View details with 'oc describe <resource>/<name>' or list everything with 'oc 
get all'. 
[root@ose-ha-master-01 playbook-openshift]# oc get SecurityContextConstraints 
privileged -o json | jq '.users' 
[ 
"system:serviceaccount:default:registry", 
"system:serviceaccount:default:router", 
"system:serviceaccount:openshift-infra:build-controller", 
"system:serviceaccount:management-infra:management-admin" 
] 
[root@ose-ha-master-01 playbook-openshift]# oadm registry \ 
> --service-account=registry \ 
> --config=/etc/origin/master/admin.kubeconfig \ 
> --credentials=/etc/origin/master/openshift-registry.kubeconfig \ 
> --images='registry.access.redhat.com/openshift3/ose-${component}:${version}' 
> \ 
> --selector="region=infra" 
DeploymentConfig "docker-registry" created 
Service "docker-registry" created 

Other than creating the dc and the svc nothing seems to happen. 

[root@ose-ha-master-01 playbook-openshift]# oc get deploymentconfig 
--all-namespaces 
NAMESPACE NAME TRIGGERS LATEST 
default docker-registry ConfigChange 1 
[root@ose-ha-master-01 playbook-openshift]# oc get pods --all-namespaces 
[root@ose-ha-master-01 playbook-openshift]# oc get svc 
NAME CLUSTER_IP EXTERNAL_IP PORT(S) SELECTOR AGE 
docker-registry 172.30.151.27 <none> 5000/TCP docker-registry=default 1m 
kubernetes 172.30.0.1 <none> 443/TCP,53/UDP,53/TCP <none> 16m 

There are no pods and this `ansible nodes -m command -a 'docker ps'` tells me 
there are no containers present anywhere. 

The nodes are schedulable. What else can I look for? 

[root@ose-ha-master-01 playbook-openshift]# oc get nodes 
NAME LABELS STATUS AGE 
ose-ha-master-01.example.com 
kubernetes.io/hostname=ose-ha-master-01.example.com,region=infra,zone=rhev 
Ready,SchedulingDisabled 28m 
ose-ha-master-02.example.com 
kubernetes.io/hostname=ose-ha-master-02.example.com,region=infra,zone=rhev 
Ready,SchedulingDisabled 28m 
ose-ha-master-03.example.com 
kubernetes.io/hostname=ose-ha-master-03.example.com,region=infra,zone=rhev 
Ready,SchedulingDisabled 28m 
ose-ha-node-01.example.com 
kubernetes.io/hostname=ose-ha-node-01.example.com,region=infra,zone=rhev Ready 
28m 
ose-ha-node-02.example.com 
kubernetes.io/hostname=ose-ha-node-02.example.com,region=infra,zone=rhev Ready 
28m 
ose-ha-node-03.example.com 
kubernetes.io/hostname=ose-ha-node-03.example.com,region=primary,zone=rhev 
Ready 28m 
ose-ha-node-04.example.com 
kubernetes.io/hostname=ose-ha-node-04.example.com,region=primary,zone=rhev 
Ready 28m 
ose-ha-node-05.example.com 
kubernetes.io/hostname=ose-ha-node-05.example.com,region=primary,zone=rhev 
Ready 28m 
ose-ha-node-06.example.com 
kubernetes.io/hostname=ose-ha-node-06.example.com,region=primary,zone=rhev 
Ready 28m 

----- On Jan 31, 2016, at 1:50 PM, Andrew Butcher <[email protected]> wrote: 

> Hey Dale,
> Two services are started when using the native ha method;
> atomic-openshift-master-api and atomic-openshift-master-controllers. How did
> you deploy the registry? Are there any failed pods?

> On Sun, Jan 31, 2016 at 4:14 PM, Dale Bewley < [email protected] > wrote:

>> I'm provisioning a Native HA OpenShift Enterprise 3.1 cluster using byo
>> playbook. There are no failures, but at the end the master service is 
>> disabled.
>> (Full hosts file here: https://gist.github.com/dlbewley/d7db07edb7fa6da72259 
>> )
>> [masters]
>> ose-ha-master-[01:03]. example.com
>> [etcd]
>> ose-ha-etcd-[01:03]. example.com
>> [lb]
>> ose-ha-lb-01.example.com
>> [nodes]
>> ose-ha-master-[01:03]. example.com openshift_node_labels="{'region': 'infra',
>> 'zone': 'rhev'}" openshift_schedulable=False
>> ose-ha-node-[01:02]. example.com openshift_node_labels="{'region': 'infra',
>> 'zone': 'rhev'}"
>> ose-ha-node-[03:06]. example.com openshift_node_labels="{'region': 'primary',
>> 'zone': 'rhev'}"

>> The playbook runs with no failures, `oc get nodes` and everything looks fine,
>> and I can even login to the web console by way of the loadbalancer node with 
>> my
>> LDAP credentials. Quite an impressive playbook. :)
>> PLAY RECAP 
>> ********************************************************************
>> localhost : ok=18 changed=0 unreachable=0 failed=0
>> ose-ha-etcd-01.example.com : ok=191 changed=40 unreachable=0 failed=0
>> ose-ha-etcd-02.example.com : ok=89 changed=20 unreachable=0 failed=0
>> ose-ha-etcd-03.example.com : ok=89 changed=20 unreachable=0 failed=0
>> ose-ha-lb-01.example.com : ok=29 changed=7 unreachable=0 failed=0
>> ose-ha-master-01.example.com : ok=334 changed=73 unreachable=0 failed=0
>> ose-ha-master-02.example.com : ok=203 changed=47 unreachable=0 failed=0
>> ose-ha-master-03.example.com : ok=203 changed=47 unreachable=0 failed=0
>> ose-ha-node-01.example.com : ok=103 changed=24 unreachable=0 failed=0
>> ose-ha-node-02.example.com : ok=103 changed=24 unreachable=0 failed=0
>> ose-ha-node-03.example.com : ok=103 changed=24 unreachable=0 failed=0
>> ose-ha-node-04.example.com : ok=103 changed=24 unreachable=0 failed=0
>> ose-ha-node-05.example.com : ok=103 changed=24 unreachable=0 failed=0
>> ose-ha-node-06.example.com : ok=103 changed=24 unreachable=0 failed=0

>> However when I attempted to deploy the registry nothing happened. I then 
>> noticed
>> that `atomic-openshift-master` service was never enabled by the playbook.

>> It looks like
>> https://github.com/openshift/openshift-ansible/blob/master/roles/openshift_master/tasks/main.yml#L274
>> is skipped because this conditional: `when: not openshift_master_ha | bool`.

>> [root@ose-ha-master-01 ]# grep -A1 'Start and enable master'
>> ansible-byo-2016-01-28.log
>> TASK: [openshift_master | Start and enable master] 
>> ****************************
>> skipping: [ ose-ha-master-01.example.com ]
>> --
>> TASK: [openshift_master | Start and enable master api] 
>> ************************
>> changed: [ ose-ha-master-01.example.com ]
>> --
>> TASK: [openshift_master | Start and enable master controller] 
>> *****************
>> changed: [ ose-ha-master-01.example.com ]
>> --
>> TASK: [openshift_master | Start and enable master] 
>> ****************************
>> skipping: [ ose-ha-master-02.example.com ]
>> --
>> TASK: [openshift_master | Start and enable master api] 
>> ************************
>> changed: [ ose-ha-master-02.example.com ]
>> --
>> TASK: [openshift_master | Start and enable master controller] 
>> *****************
>> changed: [ ose-ha-master-02.example.com ]
>> --
>> TASK: [openshift_master | Start and enable master] 
>> ****************************
>> skipping: [ ose-ha-master-03.example.com ]
>> --
>> TASK: [openshift_master | Start and enable master api] 
>> ************************
>> changed: [ ose-ha-master-03.example.com ]
>> --
>> TASK: [openshift_master | Start and enable master controller] 
>> *****************
>> changed: [ ose-ha-master-03.example.com ]

>> Any idea why this is happening? As near as I can discern openshift_master_ha 
>> is
>> true, because `"{{ groups.oo_masters_to_config | length > 1 }}"` and I see 
>> my 3
>> masters in that group.

>> _______________________________________________
>> users mailing list
>> [email protected]
>> http://lists.openshift.redhat.com/openshiftmm/listinfo/users
_______________________________________________
users mailing list
[email protected]
http://lists.openshift.redhat.com/openshiftmm/listinfo/users

Reply via email to