Hi Hetz,

You need to do one of the 2 options:

1. Enable scheduling on your master node.
2. Label the node1 and node2 with region=infra.

I would choose the second option and remove the label from the master node.



Rodrigo Bersa

Cloud Consultant, RHCSA, RHCVA

Red Hat Brasil <https://www.redhat.com>

[email protected]    M: +55 11 99557-5841 <+55-11-99557-5841>
<https://red.ht/sig> [image: Red Hat] <http://www.redhat.com.br>
TRIED. TESTED. TRUSTED. <https://redhat.com/trusted>


<http://www.redhat.com.br>

On Tue, May 23, 2017 at 10:52 AM, Ben Parees <[email protected]> wrote:

>
>
> On Tue, May 23, 2017 at 9:49 AM, Hetz Ben Hamo <[email protected]> wrote:
>
>> That's true. I didn't want to have container apps on it.
>>
>
> since you labeled it infra (based on your inventory), it won't.  but it's
> also your only infra structure labeled node, and the registry has to run on
> an infra structure node.
>
> So you either need to add another node labeled infra that's scheduleable,
> or make your master scheduleable.
>
>
>
>>
>> # oc get nodes
>> NAME          STATUS                     AGE
>> master-home   Ready,SchedulingDisabled   1h
>> node1-home    Ready                      1h
>> node2-home    Ready                      1h
>>
>>
>> תודה,
>> *חץ בן חמו*
>> אתם מוזמנים לבקר בבלוג היעוץ <http://linvirtstor.net/> או בבלוג הפרטי שלי
>> <http://benhamo.org>
>>
>> On Tue, May 23, 2017 at 4:45 PM, Ben Parees <[email protected]> wrote:
>>
>>> sounds like maybe your master node is not scheduleable, can you run:
>>>
>>> $ oc get nodes
>>>
>>> $ oc describe node master
>>>
>>> ?
>>>
>>>
>>> On Tue, May 23, 2017 at 9:42 AM, Hetz Ben Hamo <[email protected]> wrote:
>>>
>>>> Sure, here it is:
>>>>
>>>> # oc describe pod docker-registry-2-deploy
>>>> Name:                   docker-registry-2-deploy
>>>> Namespace:              default
>>>> Security Policy:        restricted
>>>> Node:                   /
>>>> Labels:                 openshift.io/deployer-pod-for.
>>>> name=docker-registry-2
>>>> Status:                 Pending
>>>> IP:
>>>> Controllers:            <none>
>>>> Containers:
>>>>   deployment:
>>>>     Image:      openshift/origin-deployer:v1.4.1
>>>>     Port:
>>>>     Volume Mounts:
>>>>       /var/run/secrets/kubernetes.io/serviceaccount from
>>>> deployer-token-sbvm4 (ro)
>>>>     Environment Variables:
>>>>       KUBERNETES_MASTER:        https://master-home:8443
>>>>       OPENSHIFT_MASTER:         https://master-home:8443
>>>>       BEARER_TOKEN_FILE:        /var/run/secrets/kubernetes.i
>>>> o/serviceaccount/token
>>>>       OPENSHIFT_CA_DATA:        -----BEGIN CERTIFICATE-----
>>>> MIIC6jCCAdKgAwIBAgIBATANBgkqhkiG9w0BAQsFADAmMSQwIgYDVQQDDBtvcGVu
>>>> c2hpZnQtc2lnbmVyQDE0OTU1NDE2MTEwHhcNMTcwNTIzMTIxMzMwWhcNMjIwNTIy
>>>> MTIxMzMxWjAmMSQwIgYDVQQDDBtvcGVuc2hpZnQtc2lnbmVyQDE0OTU1NDE2MTEw
>>>> ggEiMA0GCSqGSIb3DQEBAQUAA4IBDwAwggEKAoIBAQC9imtjAe8JBjpD99nt3D4h
>>>> VCwlWKCMugpIGWYdnHaICBS71KuIim8pWaOWYPUb73QhoUUZhZ80MYOzlB7lk/xK
>>>> NWUnQBDFYc9zKqXkxjiWlTXHv1UCyB56mxFdfxPTHN61JbE8dD9jbiBLRudgb1cq
>>>> Vhff4CRXqkdDURk8KjpnGkWW57Ky0Icp0rbOrRT/OhYv5CB8sqJedSC2VKfe9qtz
>>>> +L4ykOOa4Q1qfqD7YqPDAqnUEJFXEbqjFCdLe6q2TS0vscx/rRJcANmzApgw4BRd
>>>> OxEHH1KX6ariXSNkSWxhQIBa8qDukDrGc2dvAoLHi8ALBbnpGLE0zwtf087zdyF/
>>>> AgMBAAGjIzAhMA4GA1UdDwEB/wQEAwICpDAPBgNVHRMBAf8EBTADAQH/MA0GCSqG
>>>> SIb3DQEBCwUAA4IBAQA7Nn3iGUVH0HJN6WxR6oirpIv9VdqRgugqOoBM8O5GlV7D
>>>> 7kd4VGFSzFtXKr0kHCgA+/6sEiu0ZlQZT7IvwDWgiY/bXOY/gT8whMWVLXbXBGGT
>>>> 4brdqSRQVdgjv56kBG/cqIWedwNzItGFb+eye+AjHi20fUuVKW49Z7lvStHcvOHK
>>>> c4XyP+e3S/wg6VEMT64kAuZUvvRLhUvJK9ZfxlEGZnjQ+qrYCEpGGDjeDTeXOxMi
>>>> 6NL7Rh09p/yjemw3u+EZkfNlBMgBsA2+zEOxKbAGmENjjctFGRTJVGKq+FWR2HMi
>>>> P2pHCOPEcn2on3GAyTncdp1ANcBNTjb8gTnsoPbc
>>>> -----END CERTIFICATE-----
>>>>
>>>>       OPENSHIFT_DEPLOYMENT_NAME:        docker-registry-2
>>>>       OPENSHIFT_DEPLOYMENT_NAMESPACE:   default
>>>> Conditions:
>>>>   Type          Status
>>>>   PodScheduled  False
>>>> Volumes:
>>>>   deployer-token-sbvm4:
>>>>     Type:       Secret (a volume populated by a Secret)
>>>>     SecretName: deployer-token-sbvm4
>>>> QoS Class:      BestEffort
>>>> Tolerations:    <none>
>>>> Events:
>>>>   FirstSeen     LastSeen        Count   From
>>>>  SubobjectPath   Type            Reason                  Message
>>>>   ---------     --------        -----   ----
>>>>  -------------   --------        ------                  -------
>>>>   11m           7m              4       {default-scheduler }
>>>>          Warning         FailedScheduling        pod
>>>> (docker-registry-2-deploy) failed to fit in any node
>>>> fit failure on node (node2-home): CheckServiceAffinity,
>>>> MatchNodeSelector
>>>> fit failure on node (node1-home): MatchNodeSelector,
>>>> CheckServiceAffinity
>>>>
>>>>   4m    3m      2       {default-scheduler }            Warning
>>>> FailedScheduling        pod (docker-registry-2-deploy) failed to fit in any
>>>> node
>>>> fit failure on node (node2-home): CheckServiceAffinity,
>>>> MatchNodeSelector
>>>> fit failure on node (node1-home): CheckServiceAffinity,
>>>> MatchNodeSelector
>>>>
>>>>   11m   2m      3       {default-scheduler }            Warning
>>>> FailedScheduling        pod (docker-registry-2-deploy) failed to fit in any
>>>> node
>>>> fit failure on node (node2-home): MatchNodeSelector,
>>>> CheckServiceAffinity
>>>> fit failure on node (node1-home): CheckServiceAffinity,
>>>> MatchNodeSelector
>>>>
>>>>   13m   2m      13      {default-scheduler }            Warning
>>>> FailedScheduling        pod (docker-registry-2-deploy) failed to fit in any
>>>> node
>>>> fit failure on node (node1-home): MatchNodeSelector,
>>>> CheckServiceAffinity
>>>> fit failure on node (node2-home): MatchNodeSelector,
>>>> CheckServiceAffinity
>>>>
>>>>   11m   1m      7       {default-scheduler }            Warning
>>>> FailedScheduling        pod (docker-registry-2-deploy) failed to fit in any
>>>> node
>>>> fit failure on node (node1-home): CheckServiceAffinity,
>>>> MatchNodeSelector
>>>> fit failure on node (node2-home): MatchNodeSelector,
>>>> CheckServiceAffinity
>>>>
>>>>   13m   1m      4       {default-scheduler }            Warning
>>>> FailedScheduling        pod (docker-registry-2-deploy) failed to fit in any
>>>> node
>>>> fit failure on node (node2-home): MatchNodeSelector,
>>>> CheckServiceAffinity
>>>> fit failure on node (node1-home): MatchNodeSelector,
>>>> CheckServiceAffinity
>>>>
>>>>   13m   37s     10      {default-scheduler }            Warning
>>>> FailedScheduling        pod (docker-registry-2-deploy) failed to fit in any
>>>> node
>>>> fit failure on node (node1-home): MatchNodeSelector,
>>>> CheckServiceAffinity
>>>> fit failure on node (node2-home): CheckServiceAffinity,
>>>> MatchNodeSelector
>>>>
>>>>   13m   5s      7       {default-scheduler }            Warning
>>>> FailedScheduling        pod (docker-registry-2-deploy) failed to fit in any
>>>> node
>>>> fit failure on node (node1-home): CheckServiceAffinity,
>>>> MatchNodeSelector
>>>> fit failure on node (node2-home): CheckServiceAffinity,
>>>> MatchNodeSelector
>>>>
>>>>
>>>> # oc describe pod router-2-deploy
>>>> Name:                   router-2-deploy
>>>> Namespace:              default
>>>> Security Policy:        restricted
>>>> Node:                   /
>>>> Labels:                 openshift.io/deployer-pod-for.name=router-2
>>>> Status:                 Pending
>>>> IP:
>>>> Controllers:            <none>
>>>> Containers:
>>>>   deployment:
>>>>     Image:      openshift/origin-deployer:v1.4.1
>>>>     Port:
>>>>     Volume Mounts:
>>>>       /var/run/secrets/kubernetes.io/serviceaccount from
>>>> deployer-token-sbvm4 (ro)
>>>>     Environment Variables:
>>>>       KUBERNETES_MASTER:        https://master-home:8443
>>>>       OPENSHIFT_MASTER:         https://master-home:8443
>>>>       BEARER_TOKEN_FILE:        /var/run/secrets/kubernetes.i
>>>> o/serviceaccount/token
>>>>       OPENSHIFT_CA_DATA:        -----BEGIN CERTIFICATE-----
>>>> MIIC6jCCAdKgAwIBAgIBATANBgkqhkiG9w0BAQsFADAmMSQwIgYDVQQDDBtvcGVu
>>>> c2hpZnQtc2lnbmVyQDE0OTU1NDE2MTEwHhcNMTcwNTIzMTIxMzMwWhcNMjIwNTIy
>>>> MTIxMzMxWjAmMSQwIgYDVQQDDBtvcGVuc2hpZnQtc2lnbmVyQDE0OTU1NDE2MTEw
>>>> ggEiMA0GCSqGSIb3DQEBAQUAA4IBDwAwggEKAoIBAQC9imtjAe8JBjpD99nt3D4h
>>>> VCwlWKCMugpIGWYdnHaICBS71KuIim8pWaOWYPUb73QhoUUZhZ80MYOzlB7lk/xK
>>>> NWUnQBDFYc9zKqXkxjiWlTXHv1UCyB56mxFdfxPTHN61JbE8dD9jbiBLRudgb1cq
>>>> Vhff4CRXqkdDURk8KjpnGkWW57Ky0Icp0rbOrRT/OhYv5CB8sqJedSC2VKfe9qtz
>>>> +L4ykOOa4Q1qfqD7YqPDAqnUEJFXEbqjFCdLe6q2TS0vscx/rRJcANmzApgw4BRd
>>>> OxEHH1KX6ariXSNkSWxhQIBa8qDukDrGc2dvAoLHi8ALBbnpGLE0zwtf087zdyF/
>>>> AgMBAAGjIzAhMA4GA1UdDwEB/wQEAwICpDAPBgNVHRMBAf8EBTADAQH/MA0GCSqG
>>>> SIb3DQEBCwUAA4IBAQA7Nn3iGUVH0HJN6WxR6oirpIv9VdqRgugqOoBM8O5GlV7D
>>>> 7kd4VGFSzFtXKr0kHCgA+/6sEiu0ZlQZT7IvwDWgiY/bXOY/gT8whMWVLXbXBGGT
>>>> 4brdqSRQVdgjv56kBG/cqIWedwNzItGFb+eye+AjHi20fUuVKW49Z7lvStHcvOHK
>>>> c4XyP+e3S/wg6VEMT64kAuZUvvRLhUvJK9ZfxlEGZnjQ+qrYCEpGGDjeDTeXOxMi
>>>> 6NL7Rh09p/yjemw3u+EZkfNlBMgBsA2+zEOxKbAGmENjjctFGRTJVGKq+FWR2HMi
>>>> P2pHCOPEcn2on3GAyTncdp1ANcBNTjb8gTnsoPbc
>>>> -----END CERTIFICATE-----
>>>>
>>>>       OPENSHIFT_DEPLOYMENT_NAME:        router-2
>>>>       OPENSHIFT_DEPLOYMENT_NAMESPACE:   default
>>>> Conditions:
>>>>   Type          Status
>>>>   PodScheduled  False
>>>> Volumes:
>>>>   deployer-token-sbvm4:
>>>>     Type:       Secret (a volume populated by a Secret)
>>>>     SecretName: deployer-token-sbvm4
>>>> QoS Class:      BestEffort
>>>> Tolerations:    <none>
>>>> Events:
>>>>   FirstSeen     LastSeen        Count   From
>>>>  SubobjectPath   Type            Reason                  Message
>>>>   ---------     --------        -----   ----
>>>>  -------------   --------        ------                  -------
>>>>   13m           13m             1       {default-scheduler }
>>>>          Warning         FailedScheduling        pod (router-2-deploy)
>>>> failed to fit in any node
>>>> fit failure on node (node2-home): MatchNodeSelector,
>>>> CheckServiceAffinity
>>>> fit failure on node (node1-home): CheckServiceAffinity,
>>>> MatchNodeSelector
>>>>
>>>>   14m   10m     2       {default-scheduler }            Warning
>>>> FailedScheduling        pod (router-2-deploy) failed to fit in any node
>>>> fit failure on node (node2-home): CheckServiceAffinity,
>>>> MatchNodeSelector
>>>> fit failure on node (node1-home): MatchNodeSelector,
>>>> CheckServiceAffinity
>>>>
>>>>   15m   5m      12      {default-scheduler }            Warning
>>>> FailedScheduling        pod (router-2-deploy) failed to fit in any node
>>>> fit failure on node (node1-home): CheckServiceAffinity,
>>>> MatchNodeSelector
>>>> fit failure on node (node2-home): CheckServiceAffinity,
>>>> MatchNodeSelector
>>>>
>>>>   11m   4m      3       {default-scheduler }            Warning
>>>> FailedScheduling        pod (router-2-deploy) failed to fit in any node
>>>> fit failure on node (node2-home): CheckServiceAffinity,
>>>> MatchNodeSelector
>>>> fit failure on node (node1-home): CheckServiceAffinity,
>>>> MatchNodeSelector
>>>>
>>>>   14m   1m      10      {default-scheduler }            Warning
>>>> FailedScheduling        pod (router-2-deploy) failed to fit in any node
>>>> fit failure on node (node1-home): CheckServiceAffinity,
>>>> MatchNodeSelector
>>>> fit failure on node (node2-home): MatchNodeSelector,
>>>> CheckServiceAffinity
>>>>
>>>>   15m   54s     12      {default-scheduler }            Warning
>>>> FailedScheduling        pod (router-2-deploy) failed to fit in any node
>>>> fit failure on node (node1-home): MatchNodeSelector,
>>>> CheckServiceAffinity
>>>> fit failure on node (node2-home): MatchNodeSelector,
>>>> CheckServiceAffinity
>>>>
>>>>   15m   46s     11      {default-scheduler }            Warning
>>>> FailedScheduling        pod (router-2-deploy) failed to fit in any node
>>>> fit failure on node (node1-home): MatchNodeSelector,
>>>> CheckServiceAffinity
>>>> fit failure on node (node2-home): CheckServiceAffinity,
>>>> MatchNodeSelector
>>>>
>>>>   15m   30s     5       {default-scheduler }            Warning
>>>> FailedScheduling        pod (router-2-deploy) failed to fit in any node
>>>> fit failure on node (node2-home): MatchNodeSelector,
>>>> CheckServiceAffinity
>>>> fit failure on node (node1-home): MatchNodeSelector,
>>>> CheckServiceAffinity
>>>>
>>>> I think that there's something wrong with my ansible host file - here
>>>> it is, specially the last few lines:
>>>>
>>>> [OSEv3:children]
>>>> masters
>>>> nodes
>>>>
>>>> [OSEv3:vars]
>>>> ansible_ssh_user=root
>>>> # ansible_become=yes
>>>>
>>>> deployment_type=origin
>>>> openshift_release=v1.4
>>>> openshift_image_tag=v1.4.1
>>>> containerized=true
>>>> openshift_install_examples=true
>>>> # openshift_hosted_metrics_deploy=true
>>>>
>>>> # use htpasswd authentication with demo/demo
>>>> openshift_master_identity_providers=[{'name': 'htpasswd_auth',
>>>> 'login': 'true', 'challenge': 'true', 'kind': 
>>>> 'HTPasswdPasswordIdentityProvider',
>>>> 'filename': '/etc/origin/master/htpasswd'}]
>>>> openshift_master_htpasswd_users={'demo': '$apr1$.MaA77kd$Rlnn6RXq9kCjnE
>>>> fh5I3w/.'}
>>>>
>>>> # put the router on dedicated infra node
>>>> openshift_hosted_router_selector='region=infra'
>>>> openshift_master_default_subdomain=apps.hetzlabs.pro
>>>>
>>>> # put the image registry on dedicated infra node
>>>> openshift_hosted_registry_selector='region=infra'
>>>>
>>>> # project pods should be placed on primary nodes
>>>> osm_default_node_selector='region=primary'
>>>>
>>>> [masters]
>>>> master-home.hetzlabs.pro openshift_public_hostname="mas
>>>> ter-home.hetzlabs.pro"
>>>>
>>>> [nodes]
>>>> # master needs to be included in the node to be configured in the SDN
>>>> master-home.hetzlabs.pro openshift_node_labels="{'region': 'infra',
>>>> 'zone': 'default'}"
>>>> node1-home.hetzlabs.pro openshift_node_labels="{'region': 'primary',
>>>> 'zone': 'default'}"
>>>> node2-home.hetzlabs.pro openshift_node_labels="{'region': 'primary',
>>>> 'zone': 'default'}"
>>>>
>>>> Basically I'm looking to run openshift origin (1.4 or 1.5) with 1
>>>> master, 2 nodes (total: 3 vm's). Am I doing it right?
>>>>
>>>> Thanks
>>>>
>>>>
>>>>
>>>> On Tue, May 23, 2017 at 4:24 PM, Rodrigo Bersa <[email protected]>
>>>> wrote:
>>>>
>>>>> Hi Hetz,
>>>>>
>>>>> It seems that your Registry and Router PODs are not running. Probably
>>>>> there's a problem avoiding them to deploy.
>>>>>
>>>>> Can you send the output of the commands below?
>>>>>
>>>>> # oc describe pod docker-registry-1-deploy
>>>>> # oc describe pod router-1-deploy
>>>>>
>>>>>
>>>>>
>>>>> Rodrigo Bersa
>>>>>
>>>>> Cloud Consultant, RHCSA, RHCVA
>>>>>
>>>>> Red Hat Brasil <https://www.redhat.com>
>>>>>
>>>>> [email protected]    M: +55 11 99557-5841 <+55-11-99557-5841>
>>>>> <https://red.ht/sig> [image: Red Hat] <http://www.redhat.com.br>
>>>>> TRIED. TESTED. TRUSTED. <https://redhat.com/trusted>
>>>>>
>>>>>
>>>>> <http://www.redhat.com.br>
>>>>>
>>>>> On Tue, May 23, 2017 at 8:28 AM, Hetz Ben Hamo <[email protected]> wrote:
>>>>>
>>>>>> ]# oc get pods -n default
>>>>>> NAME                        READY     STATUS    RESTARTS   AGE
>>>>>> docker-registry-1-deploy    0/1       Pending   0          16m
>>>>>> registry-console-1-deploy   0/1       Error     0          15m
>>>>>> router-1-deploy             0/1       Pending   0          17m
>>>>>> [root@master-home ~]# oc logs registry-console-1-deploy
>>>>>> --> Scaling registry-console-1 to 1
>>>>>> --> Waiting up to 10m0s for pods in rc registry-console-1 to become
>>>>>> ready
>>>>>> error: update acceptor rejected registry-console-1: pods for rc
>>>>>> "registry-console-1" took longer than 600 seconds to become ready
>>>>>> [root@master-home ~]# oc logs router-1-deploy
>>>>>> [root@master-home ~]# oc logs docker-registry-1-deploy
>>>>>> [root@master-home ~]# oc logs docker-registry-1-deploy -n default
>>>>>> [root@master-home ~]# oc get pods
>>>>>>
>>>>>>
>>>>>> תודה,
>>>>>> *חץ בן חמו*
>>>>>> אתם מוזמנים לבקר בבלוג היעוץ <http://linvirtstor.net/> או בבלוג הפרטי
>>>>>> שלי <http://benhamo.org>
>>>>>>
>>>>>> On Tue, May 23, 2017 at 1:49 AM, Ben Parees <[email protected]>
>>>>>> wrote:
>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> On Mon, May 22, 2017 at 6:18 PM, Hetz Ben Hamo <[email protected]>
>>>>>>> wrote:
>>>>>>>
>>>>>>>> Hi,
>>>>>>>>
>>>>>>>> I've built on a 3 nodes openshift origin using the host file
>>>>>>>> included below, but it seems few things are getting broken. I didn't 
>>>>>>>> modify
>>>>>>>> anything yet on the openshift, just used the openshift-Ansible checked 
>>>>>>>> out
>>>>>>>> from today.
>>>>>>>>
>>>>>>>> Problem one: After building an image from the examples (I chose
>>>>>>>> Java with the example of wildfly) I get:
>>>>>>>>
>>>>>>>> [INFO] ------------------------------------------------------------
>>>>>>>> ------------
>>>>>>>> [INFO] BUILD SUCCESS
>>>>>>>> [INFO] ------------------------------------------------------------
>>>>>>>> ------------
>>>>>>>> [INFO] Total time: 12.182 s
>>>>>>>> [INFO] Finished at: 2017-05-22T22:08:21+00:00
>>>>>>>> [INFO] Final Memory: 14M/134M
>>>>>>>> [INFO] ------------------------------------------------------------
>>>>>>>> ------------
>>>>>>>> Moving built war files into /wildfly/standalone/deployments for
>>>>>>>> later deployment...
>>>>>>>> Moving all war artifacts from /opt/app-root/src/target directory
>>>>>>>> into /wildfly/standalone/deployments for later deployment...
>>>>>>>> '/opt/app-root/src/target/ROOT.war' ->
>>>>>>>> '/wildfly/standalone/deployments/ROOT.war'
>>>>>>>> Moving all ear artifacts from /opt/app-root/src/target directory
>>>>>>>> into /wildfly/standalone/deployments for later deployment...
>>>>>>>> Moving all rar artifacts from /opt/app-root/src/target directory
>>>>>>>> into /wildfly/standalone/deployments for later deployment...
>>>>>>>> Moving all jar artifacts from /opt/app-root/src/target directory
>>>>>>>> into /wildfly/standalone/deployments for later deployment...
>>>>>>>> ...done
>>>>>>>> Pushing image 172.30.172.85:5000/test1/wf:latest ...
>>>>>>>> Warning: Push failed, retrying in 5s ...
>>>>>>>> Warning: Push failed, retrying in 5s ...
>>>>>>>> Warning: Push failed, retrying in 5s ...
>>>>>>>> Warning: Push failed, retrying in 5s ...
>>>>>>>> Warning: Push failed, retrying in 5s ...
>>>>>>>> Warning: Push failed, retrying in 5s ...
>>>>>>>> Warning: Push failed, retrying in 5s ...
>>>>>>>> Registry server Address:
>>>>>>>> Registry server User Name: serviceaccount
>>>>>>>> Registry server Email: [email protected]
>>>>>>>> Registry server Password: <<non-empty>>
>>>>>>>> error: build error: Failed to push image: Get
>>>>>>>> https://172.30.172.85:5000/v1/_ping: dial tcp 172.30.172.85:5000:
>>>>>>>> getsockopt: connection refused
>>>>>>>>
>>>>>>>>
>>>>>>> can you confirm your registry pod is running in the default
>>>>>>> namespace (oc get pods -n default)?  Can you get logs from it?
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>>
>>>>>>>> Another problem: I added the metrics option so it installed hawkler
>>>>>>>> but when it complains that it needs SSL approval (it shows a message 
>>>>>>>> about
>>>>>>>> a problem with hawkler and gives a link to open it) I get upon 
>>>>>>>> clicking the
>>>>>>>> link: connection refused.
>>>>>>>>
>>>>>>>> I've tested the host configuration on 2 sets of VM's (one at home
>>>>>>>> with digital ocean, another set here at home with VMWare). I've set up 
>>>>>>>> DNS
>>>>>>>> with subdomain wildcard and I can ping the app names but trying to 
>>>>>>>> connect
>>>>>>>> through a browser or curl - gives connection refused.
>>>>>>>>
>>>>>>>> Have I missed something?
>>>>>>>>
>>>>>>>> here is my byo host file:
>>>>>>>>
>>>>>>>> [OSEv3:children]
>>>>>>>> masters
>>>>>>>> nodes
>>>>>>>>
>>>>>>>> [OSEv3:vars]
>>>>>>>> ansible_ssh_user=root
>>>>>>>>
>>>>>>>> deployment_type=origin
>>>>>>>> openshift_release=v1.5.0
>>>>>>>> containerized=true
>>>>>>>> openshift_install_examples=true
>>>>>>>> openshift_hosted_metrics_deploy=true
>>>>>>>>
>>>>>>>> # use htpasswd authentication with demo/demo
>>>>>>>> openshift_master_identity.providers=[{'name': 'htpasswd_auth',
>>>>>>>> 'login': 'true', 'challenge': 'true', 'kind': 
>>>>>>>> 'HTPasswdPasswordIdentityProvider',
>>>>>>>> 'filename': '/etc/origin/master/htpasswd'}]
>>>>>>>> openshift_master_htpasswd_users={'demo':
>>>>>>>> '$XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX.'}
>>>>>>>>
>>>>>>>> # put the router on dedicated infra node
>>>>>>>> openshift_hosted_router_selector='region=infra'
>>>>>>>> openshift_master_default_subdomain=apps.test.com
>>>>>>>>
>>>>>>>> # put the image registry on dedicated infra node
>>>>>>>> openshift_hosted_registry_selector='region=infra'
>>>>>>>>
>>>>>>>> #.project pods should be placed on primary nodes
>>>>>>>> osm_default_node_selector='region=primary'
>>>>>>>>
>>>>>>>> [masters]
>>>>>>>> master-home.test.com openshift_public_hostname="mas
>>>>>>>> ter-home.test.com"
>>>>>>>>
>>>>>>>> [nodes]
>>>>>>>> # master needs to be included in the node to be configured in the
>>>>>>>> SDN
>>>>>>>> # master-home.test.com
>>>>>>>> master-home.test.com openshift_node_labels="{'region': 'infra',
>>>>>>>> 'zone': 'default'}"
>>>>>>>> node1-home.test.com openshift_node_labels="{'region': 'primary',
>>>>>>>> 'zone': 'default'}"
>>>>>>>> node2-home.test.com openshift_node_labels="{'region': 'primary',
>>>>>>>> 'zone': 'default'}"
>>>>>>>>
>>>>>>>>
>>>>>>>> Thanks
>>>>>>>>
>>>>>>>> _______________________________________________
>>>>>>>> users mailing list
>>>>>>>> [email protected]
>>>>>>>> http://lists.openshift.redhat.com/openshiftmm/listinfo/users
>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> --
>>>>>>> Ben Parees | OpenShift
>>>>>>>
>>>>>>>
>>>>>>
>>>>>> _______________________________________________
>>>>>> users mailing list
>>>>>> [email protected]
>>>>>> http://lists.openshift.redhat.com/openshiftmm/listinfo/users
>>>>>>
>>>>>>
>>>>>
>>>>
>>>
>>>
>>> --
>>> Ben Parees | OpenShift
>>>
>>>
>>
>
>
> --
> Ben Parees | OpenShift
>
>
_______________________________________________
users mailing list
[email protected]
http://lists.openshift.redhat.com/openshiftmm/listinfo/users

Reply via email to