Sure, here it is:

# oc describe pod docker-registry-2-deploy
Name:                   docker-registry-2-deploy
Namespace:              default
Security Policy:        restricted
Node:                   /
Labels:                 openshift.io/deployer-pod-for.name=docker-registry-2
Status:                 Pending
IP:
Controllers:            <none>
Containers:
  deployment:
    Image:      openshift/origin-deployer:v1.4.1
    Port:
    Volume Mounts:
      /var/run/secrets/kubernetes.io/serviceaccount from
deployer-token-sbvm4 (ro)
    Environment Variables:
      KUBERNETES_MASTER:        https://master-home:8443
      OPENSHIFT_MASTER:         https://master-home:8443
      BEARER_TOKEN_FILE:        /var/run/secrets/
kubernetes.io/serviceaccount/token
      OPENSHIFT_CA_DATA:        -----BEGIN CERTIFICATE-----
MIIC6jCCAdKgAwIBAgIBATANBgkqhkiG9w0BAQsFADAmMSQwIgYDVQQDDBtvcGVu
c2hpZnQtc2lnbmVyQDE0OTU1NDE2MTEwHhcNMTcwNTIzMTIxMzMwWhcNMjIwNTIy
MTIxMzMxWjAmMSQwIgYDVQQDDBtvcGVuc2hpZnQtc2lnbmVyQDE0OTU1NDE2MTEw
ggEiMA0GCSqGSIb3DQEBAQUAA4IBDwAwggEKAoIBAQC9imtjAe8JBjpD99nt3D4h
VCwlWKCMugpIGWYdnHaICBS71KuIim8pWaOWYPUb73QhoUUZhZ80MYOzlB7lk/xK
NWUnQBDFYc9zKqXkxjiWlTXHv1UCyB56mxFdfxPTHN61JbE8dD9jbiBLRudgb1cq
Vhff4CRXqkdDURk8KjpnGkWW57Ky0Icp0rbOrRT/OhYv5CB8sqJedSC2VKfe9qtz
+L4ykOOa4Q1qfqD7YqPDAqnUEJFXEbqjFCdLe6q2TS0vscx/rRJcANmzApgw4BRd
OxEHH1KX6ariXSNkSWxhQIBa8qDukDrGc2dvAoLHi8ALBbnpGLE0zwtf087zdyF/
AgMBAAGjIzAhMA4GA1UdDwEB/wQEAwICpDAPBgNVHRMBAf8EBTADAQH/MA0GCSqG
SIb3DQEBCwUAA4IBAQA7Nn3iGUVH0HJN6WxR6oirpIv9VdqRgugqOoBM8O5GlV7D
7kd4VGFSzFtXKr0kHCgA+/6sEiu0ZlQZT7IvwDWgiY/bXOY/gT8whMWVLXbXBGGT
4brdqSRQVdgjv56kBG/cqIWedwNzItGFb+eye+AjHi20fUuVKW49Z7lvStHcvOHK
c4XyP+e3S/wg6VEMT64kAuZUvvRLhUvJK9ZfxlEGZnjQ+qrYCEpGGDjeDTeXOxMi
6NL7Rh09p/yjemw3u+EZkfNlBMgBsA2+zEOxKbAGmENjjctFGRTJVGKq+FWR2HMi
P2pHCOPEcn2on3GAyTncdp1ANcBNTjb8gTnsoPbc
-----END CERTIFICATE-----

      OPENSHIFT_DEPLOYMENT_NAME:        docker-registry-2
      OPENSHIFT_DEPLOYMENT_NAMESPACE:   default
Conditions:
  Type          Status
  PodScheduled  False
Volumes:
  deployer-token-sbvm4:
    Type:       Secret (a volume populated by a Secret)
    SecretName: deployer-token-sbvm4
QoS Class:      BestEffort
Tolerations:    <none>
Events:
  FirstSeen     LastSeen        Count   From
 SubobjectPath   Type            Reason                  Message
  ---------     --------        -----   ----
 -------------   --------        ------                  -------
  11m           7m              4       {default-scheduler }
     Warning         FailedScheduling        pod (docker-registry-2-deploy)
failed to fit in any node
fit failure on node (node2-home): CheckServiceAffinity, MatchNodeSelector
fit failure on node (node1-home): MatchNodeSelector, CheckServiceAffinity

  4m    3m      2       {default-scheduler }            Warning
FailedScheduling        pod (docker-registry-2-deploy) failed to fit in any
node
fit failure on node (node2-home): CheckServiceAffinity, MatchNodeSelector
fit failure on node (node1-home): CheckServiceAffinity, MatchNodeSelector

  11m   2m      3       {default-scheduler }            Warning
FailedScheduling        pod (docker-registry-2-deploy) failed to fit in any
node
fit failure on node (node2-home): MatchNodeSelector, CheckServiceAffinity
fit failure on node (node1-home): CheckServiceAffinity, MatchNodeSelector

  13m   2m      13      {default-scheduler }            Warning
FailedScheduling        pod (docker-registry-2-deploy) failed to fit in any
node
fit failure on node (node1-home): MatchNodeSelector, CheckServiceAffinity
fit failure on node (node2-home): MatchNodeSelector, CheckServiceAffinity

  11m   1m      7       {default-scheduler }            Warning
FailedScheduling        pod (docker-registry-2-deploy) failed to fit in any
node
fit failure on node (node1-home): CheckServiceAffinity, MatchNodeSelector
fit failure on node (node2-home): MatchNodeSelector, CheckServiceAffinity

  13m   1m      4       {default-scheduler }            Warning
FailedScheduling        pod (docker-registry-2-deploy) failed to fit in any
node
fit failure on node (node2-home): MatchNodeSelector, CheckServiceAffinity
fit failure on node (node1-home): MatchNodeSelector, CheckServiceAffinity

  13m   37s     10      {default-scheduler }            Warning
FailedScheduling        pod (docker-registry-2-deploy) failed to fit in any
node
fit failure on node (node1-home): MatchNodeSelector, CheckServiceAffinity
fit failure on node (node2-home): CheckServiceAffinity, MatchNodeSelector

  13m   5s      7       {default-scheduler }            Warning
FailedScheduling        pod (docker-registry-2-deploy) failed to fit in any
node
fit failure on node (node1-home): CheckServiceAffinity, MatchNodeSelector
fit failure on node (node2-home): CheckServiceAffinity, MatchNodeSelector


# oc describe pod router-2-deploy
Name:                   router-2-deploy
Namespace:              default
Security Policy:        restricted
Node:                   /
Labels:                 openshift.io/deployer-pod-for.name=router-2
Status:                 Pending
IP:
Controllers:            <none>
Containers:
  deployment:
    Image:      openshift/origin-deployer:v1.4.1
    Port:
    Volume Mounts:
      /var/run/secrets/kubernetes.io/serviceaccount from
deployer-token-sbvm4 (ro)
    Environment Variables:
      KUBERNETES_MASTER:        https://master-home:8443
      OPENSHIFT_MASTER:         https://master-home:8443
      BEARER_TOKEN_FILE:        /var/run/secrets/
kubernetes.io/serviceaccount/token
      OPENSHIFT_CA_DATA:        -----BEGIN CERTIFICATE-----
MIIC6jCCAdKgAwIBAgIBATANBgkqhkiG9w0BAQsFADAmMSQwIgYDVQQDDBtvcGVu
c2hpZnQtc2lnbmVyQDE0OTU1NDE2MTEwHhcNMTcwNTIzMTIxMzMwWhcNMjIwNTIy
MTIxMzMxWjAmMSQwIgYDVQQDDBtvcGVuc2hpZnQtc2lnbmVyQDE0OTU1NDE2MTEw
ggEiMA0GCSqGSIb3DQEBAQUAA4IBDwAwggEKAoIBAQC9imtjAe8JBjpD99nt3D4h
VCwlWKCMugpIGWYdnHaICBS71KuIim8pWaOWYPUb73QhoUUZhZ80MYOzlB7lk/xK
NWUnQBDFYc9zKqXkxjiWlTXHv1UCyB56mxFdfxPTHN61JbE8dD9jbiBLRudgb1cq
Vhff4CRXqkdDURk8KjpnGkWW57Ky0Icp0rbOrRT/OhYv5CB8sqJedSC2VKfe9qtz
+L4ykOOa4Q1qfqD7YqPDAqnUEJFXEbqjFCdLe6q2TS0vscx/rRJcANmzApgw4BRd
OxEHH1KX6ariXSNkSWxhQIBa8qDukDrGc2dvAoLHi8ALBbnpGLE0zwtf087zdyF/
AgMBAAGjIzAhMA4GA1UdDwEB/wQEAwICpDAPBgNVHRMBAf8EBTADAQH/MA0GCSqG
SIb3DQEBCwUAA4IBAQA7Nn3iGUVH0HJN6WxR6oirpIv9VdqRgugqOoBM8O5GlV7D
7kd4VGFSzFtXKr0kHCgA+/6sEiu0ZlQZT7IvwDWgiY/bXOY/gT8whMWVLXbXBGGT
4brdqSRQVdgjv56kBG/cqIWedwNzItGFb+eye+AjHi20fUuVKW49Z7lvStHcvOHK
c4XyP+e3S/wg6VEMT64kAuZUvvRLhUvJK9ZfxlEGZnjQ+qrYCEpGGDjeDTeXOxMi
6NL7Rh09p/yjemw3u+EZkfNlBMgBsA2+zEOxKbAGmENjjctFGRTJVGKq+FWR2HMi
P2pHCOPEcn2on3GAyTncdp1ANcBNTjb8gTnsoPbc
-----END CERTIFICATE-----

      OPENSHIFT_DEPLOYMENT_NAME:        router-2
      OPENSHIFT_DEPLOYMENT_NAMESPACE:   default
Conditions:
  Type          Status
  PodScheduled  False
Volumes:
  deployer-token-sbvm4:
    Type:       Secret (a volume populated by a Secret)
    SecretName: deployer-token-sbvm4
QoS Class:      BestEffort
Tolerations:    <none>
Events:
  FirstSeen     LastSeen        Count   From
 SubobjectPath   Type            Reason                  Message
  ---------     --------        -----   ----
 -------------   --------        ------                  -------
  13m           13m             1       {default-scheduler }
     Warning         FailedScheduling        pod (router-2-deploy) failed
to fit in any node
fit failure on node (node2-home): MatchNodeSelector, CheckServiceAffinity
fit failure on node (node1-home): CheckServiceAffinity, MatchNodeSelector

  14m   10m     2       {default-scheduler }            Warning
FailedScheduling        pod (router-2-deploy) failed to fit in any node
fit failure on node (node2-home): CheckServiceAffinity, MatchNodeSelector
fit failure on node (node1-home): MatchNodeSelector, CheckServiceAffinity

  15m   5m      12      {default-scheduler }            Warning
FailedScheduling        pod (router-2-deploy) failed to fit in any node
fit failure on node (node1-home): CheckServiceAffinity, MatchNodeSelector
fit failure on node (node2-home): CheckServiceAffinity, MatchNodeSelector

  11m   4m      3       {default-scheduler }            Warning
FailedScheduling        pod (router-2-deploy) failed to fit in any node
fit failure on node (node2-home): CheckServiceAffinity, MatchNodeSelector
fit failure on node (node1-home): CheckServiceAffinity, MatchNodeSelector

  14m   1m      10      {default-scheduler }            Warning
FailedScheduling        pod (router-2-deploy) failed to fit in any node
fit failure on node (node1-home): CheckServiceAffinity, MatchNodeSelector
fit failure on node (node2-home): MatchNodeSelector, CheckServiceAffinity

  15m   54s     12      {default-scheduler }            Warning
FailedScheduling        pod (router-2-deploy) failed to fit in any node
fit failure on node (node1-home): MatchNodeSelector, CheckServiceAffinity
fit failure on node (node2-home): MatchNodeSelector, CheckServiceAffinity

  15m   46s     11      {default-scheduler }            Warning
FailedScheduling        pod (router-2-deploy) failed to fit in any node
fit failure on node (node1-home): MatchNodeSelector, CheckServiceAffinity
fit failure on node (node2-home): CheckServiceAffinity, MatchNodeSelector

  15m   30s     5       {default-scheduler }            Warning
FailedScheduling        pod (router-2-deploy) failed to fit in any node
fit failure on node (node2-home): MatchNodeSelector, CheckServiceAffinity
fit failure on node (node1-home): MatchNodeSelector, CheckServiceAffinity

I think that there's something wrong with my ansible host file - here it
is, specially the last few lines:

[OSEv3:children]
masters
nodes

[OSEv3:vars]
ansible_ssh_user=root
# ansible_become=yes

deployment_type=origin
openshift_release=v1.4
openshift_image_tag=v1.4.1
containerized=true
openshift_install_examples=true
# openshift_hosted_metrics_deploy=true

# use htpasswd authentication with demo/demo
openshift_master_identity_providers=[{'name': 'htpasswd_auth', 'login':
'true', 'challenge': 'true', 'kind': 'HTPasswdPasswordIdentityProvider',
'filename': '/etc/origin/master/htpasswd'}]
openshift_master_htpasswd_users={'demo':
'$apr1$.MaA77kd$Rlnn6RXq9kCjnEfh5I3w/.'}

# put the router on dedicated infra node
openshift_hosted_router_selector='region=infra'
openshift_master_default_subdomain=apps.hetzlabs.pro

# put the image registry on dedicated infra node
openshift_hosted_registry_selector='region=infra'

# project pods should be placed on primary nodes
osm_default_node_selector='region=primary'

[masters]
master-home.hetzlabs.pro openshift_public_hostname="master-home.hetzlabs.pro
"

[nodes]
# master needs to be included in the node to be configured in the SDN
master-home.hetzlabs.pro openshift_node_labels="{'region': 'infra', 'zone':
'default'}"
node1-home.hetzlabs.pro openshift_node_labels="{'region': 'primary',
'zone': 'default'}"
node2-home.hetzlabs.pro openshift_node_labels="{'region': 'primary',
'zone': 'default'}"

Basically I'm looking to run openshift origin (1.4 or 1.5) with 1 master, 2
nodes (total: 3 vm's). Am I doing it right?

Thanks



On Tue, May 23, 2017 at 4:24 PM, Rodrigo Bersa <[email protected]> wrote:

> Hi Hetz,
>
> It seems that your Registry and Router PODs are not running. Probably
> there's a problem avoiding them to deploy.
>
> Can you send the output of the commands below?
>
> # oc describe pod docker-registry-1-deploy
> # oc describe pod router-1-deploy
>
>
>
> Rodrigo Bersa
>
> Cloud Consultant, RHCSA, RHCVA
>
> Red Hat Brasil <https://www.redhat.com>
>
> [email protected]    M: +55 11 99557-5841 <+55-11-99557-5841>
> <https://red.ht/sig> [image: Red Hat] <http://www.redhat.com.br>
> TRIED. TESTED. TRUSTED. <https://redhat.com/trusted>
>
>
> <http://www.redhat.com.br>
>
> On Tue, May 23, 2017 at 8:28 AM, Hetz Ben Hamo <[email protected]> wrote:
>
>> ]# oc get pods -n default
>> NAME                        READY     STATUS    RESTARTS   AGE
>> docker-registry-1-deploy    0/1       Pending   0          16m
>> registry-console-1-deploy   0/1       Error     0          15m
>> router-1-deploy             0/1       Pending   0          17m
>> [root@master-home ~]# oc logs registry-console-1-deploy
>> --> Scaling registry-console-1 to 1
>> --> Waiting up to 10m0s for pods in rc registry-console-1 to become ready
>> error: update acceptor rejected registry-console-1: pods for rc
>> "registry-console-1" took longer than 600 seconds to become ready
>> [root@master-home ~]# oc logs router-1-deploy
>> [root@master-home ~]# oc logs docker-registry-1-deploy
>> [root@master-home ~]# oc logs docker-registry-1-deploy -n default
>> [root@master-home ~]# oc get pods
>>
>>
>> תודה,
>> *חץ בן חמו*
>> אתם מוזמנים לבקר בבלוג היעוץ <http://linvirtstor.net/> או בבלוג הפרטי שלי
>> <http://benhamo.org>
>>
>> On Tue, May 23, 2017 at 1:49 AM, Ben Parees <[email protected]> wrote:
>>
>>>
>>>
>>> On Mon, May 22, 2017 at 6:18 PM, Hetz Ben Hamo <[email protected]> wrote:
>>>
>>>> Hi,
>>>>
>>>> I've built on a 3 nodes openshift origin using the host file included
>>>> below, but it seems few things are getting broken. I didn't modify anything
>>>> yet on the openshift, just used the openshift-Ansible checked out from
>>>> today.
>>>>
>>>> Problem one: After building an image from the examples (I chose Java
>>>> with the example of wildfly) I get:
>>>>
>>>> [INFO] ------------------------------------------------------------
>>>> ------------
>>>> [INFO] BUILD SUCCESS
>>>> [INFO] ------------------------------------------------------------
>>>> ------------
>>>> [INFO] Total time: 12.182 s
>>>> [INFO] Finished at: 2017-05-22T22:08:21+00:00
>>>> [INFO] Final Memory: 14M/134M
>>>> [INFO] ------------------------------------------------------------
>>>> ------------
>>>> Moving built war files into /wildfly/standalone/deployments for later
>>>> deployment...
>>>> Moving all war artifacts from /opt/app-root/src/target directory into
>>>> /wildfly/standalone/deployments for later deployment...
>>>> '/opt/app-root/src/target/ROOT.war' -> '/wildfly/standalone/deploymen
>>>> ts/ROOT.war'
>>>> Moving all ear artifacts from /opt/app-root/src/target directory into
>>>> /wildfly/standalone/deployments for later deployment...
>>>> Moving all rar artifacts from /opt/app-root/src/target directory into
>>>> /wildfly/standalone/deployments for later deployment...
>>>> Moving all jar artifacts from /opt/app-root/src/target directory into
>>>> /wildfly/standalone/deployments for later deployment...
>>>> ...done
>>>> Pushing image 172.30.172.85:5000/test1/wf:latest ...
>>>> Warning: Push failed, retrying in 5s ...
>>>> Warning: Push failed, retrying in 5s ...
>>>> Warning: Push failed, retrying in 5s ...
>>>> Warning: Push failed, retrying in 5s ...
>>>> Warning: Push failed, retrying in 5s ...
>>>> Warning: Push failed, retrying in 5s ...
>>>> Warning: Push failed, retrying in 5s ...
>>>> Registry server Address:
>>>> Registry server User Name: serviceaccount
>>>> Registry server Email: [email protected]
>>>> Registry server Password: <<non-empty>>
>>>> error: build error: Failed to push image: Get
>>>> https://172.30.172.85:5000/v1/_ping: dial tcp 172.30.172.85:5000:
>>>> getsockopt: connection refused
>>>>
>>>>
>>> can you confirm your registry pod is running in the default namespace
>>> (oc get pods -n default)?  Can you get logs from it?
>>>
>>>
>>>
>>>>
>>>> Another problem: I added the metrics option so it installed hawkler but
>>>> when it complains that it needs SSL approval (it shows a message about a
>>>> problem with hawkler and gives a link to open it) I get upon clicking the
>>>> link: connection refused.
>>>>
>>>> I've tested the host configuration on 2 sets of VM's (one at home with
>>>> digital ocean, another set here at home with VMWare). I've set up DNS with
>>>> subdomain wildcard and I can ping the app names but trying to connect
>>>> through a browser or curl - gives connection refused.
>>>>
>>>> Have I missed something?
>>>>
>>>> here is my byo host file:
>>>>
>>>> [OSEv3:children]
>>>> masters
>>>> nodes
>>>>
>>>> [OSEv3:vars]
>>>> ansible_ssh_user=root
>>>>
>>>> deployment_type=origin
>>>> openshift_release=v1.5.0
>>>> containerized=true
>>>> openshift_install_examples=true
>>>> openshift_hosted_metrics_deploy=true
>>>>
>>>> # use htpasswd authentication with demo/demo
>>>> openshift_master_identity.providers=[{'name': 'htpasswd_auth',
>>>> 'login': 'true', 'challenge': 'true', 'kind': 
>>>> 'HTPasswdPasswordIdentityProvider',
>>>> 'filename': '/etc/origin/master/htpasswd'}]
>>>> openshift_master_htpasswd_users={'demo': '$XXXXXXXXXXXXXXXXXXXXXXXXXXXX
>>>> XXXXXXXX.'}
>>>>
>>>> # put the router on dedicated infra node
>>>> openshift_hosted_router_selector='region=infra'
>>>> openshift_master_default_subdomain=apps.test.com
>>>>
>>>> # put the image registry on dedicated infra node
>>>> openshift_hosted_registry_selector='region=infra'
>>>>
>>>> #.project pods should be placed on primary nodes
>>>> osm_default_node_selector='region=primary'
>>>>
>>>> [masters]
>>>> master-home.test.com openshift_public_hostname="master-home.test.com"
>>>>
>>>> [nodes]
>>>> # master needs to be included in the node to be configured in the SDN
>>>> # master-home.test.com
>>>> master-home.test.com openshift_node_labels="{'region': 'infra',
>>>> 'zone': 'default'}"
>>>> node1-home.test.com openshift_node_labels="{'region': 'primary',
>>>> 'zone': 'default'}"
>>>> node2-home.test.com openshift_node_labels="{'region': 'primary',
>>>> 'zone': 'default'}"
>>>>
>>>>
>>>> Thanks
>>>>
>>>> _______________________________________________
>>>> users mailing list
>>>> [email protected]
>>>> http://lists.openshift.redhat.com/openshiftmm/listinfo/users
>>>>
>>>>
>>>
>>>
>>> --
>>> Ben Parees | OpenShift
>>>
>>>
>>
>> _______________________________________________
>> users mailing list
>> [email protected]
>> http://lists.openshift.redhat.com/openshiftmm/listinfo/users
>>
>>
>
_______________________________________________
users mailing list
[email protected]
http://lists.openshift.redhat.com/openshiftmm/listinfo/users

Reply via email to