Sure, here it is:
# oc describe pod docker-registry-2-deploy
Name: docker-registry-2-deploy
Namespace: default
Security Policy: restricted
Node: /
Labels: openshift.io/deployer-pod-for.name=docker-registry-2
Status: Pending
IP:
Controllers: <none>
Containers:
deployment:
Image: openshift/origin-deployer:v1.4.1
Port:
Volume Mounts:
/var/run/secrets/kubernetes.io/serviceaccount from
deployer-token-sbvm4 (ro)
Environment Variables:
KUBERNETES_MASTER: https://master-home:8443
OPENSHIFT_MASTER: https://master-home:8443
BEARER_TOKEN_FILE: /var/run/secrets/
kubernetes.io/serviceaccount/token
OPENSHIFT_CA_DATA: -----BEGIN CERTIFICATE-----
MIIC6jCCAdKgAwIBAgIBATANBgkqhkiG9w0BAQsFADAmMSQwIgYDVQQDDBtvcGVu
c2hpZnQtc2lnbmVyQDE0OTU1NDE2MTEwHhcNMTcwNTIzMTIxMzMwWhcNMjIwNTIy
MTIxMzMxWjAmMSQwIgYDVQQDDBtvcGVuc2hpZnQtc2lnbmVyQDE0OTU1NDE2MTEw
ggEiMA0GCSqGSIb3DQEBAQUAA4IBDwAwggEKAoIBAQC9imtjAe8JBjpD99nt3D4h
VCwlWKCMugpIGWYdnHaICBS71KuIim8pWaOWYPUb73QhoUUZhZ80MYOzlB7lk/xK
NWUnQBDFYc9zKqXkxjiWlTXHv1UCyB56mxFdfxPTHN61JbE8dD9jbiBLRudgb1cq
Vhff4CRXqkdDURk8KjpnGkWW57Ky0Icp0rbOrRT/OhYv5CB8sqJedSC2VKfe9qtz
+L4ykOOa4Q1qfqD7YqPDAqnUEJFXEbqjFCdLe6q2TS0vscx/rRJcANmzApgw4BRd
OxEHH1KX6ariXSNkSWxhQIBa8qDukDrGc2dvAoLHi8ALBbnpGLE0zwtf087zdyF/
AgMBAAGjIzAhMA4GA1UdDwEB/wQEAwICpDAPBgNVHRMBAf8EBTADAQH/MA0GCSqG
SIb3DQEBCwUAA4IBAQA7Nn3iGUVH0HJN6WxR6oirpIv9VdqRgugqOoBM8O5GlV7D
7kd4VGFSzFtXKr0kHCgA+/6sEiu0ZlQZT7IvwDWgiY/bXOY/gT8whMWVLXbXBGGT
4brdqSRQVdgjv56kBG/cqIWedwNzItGFb+eye+AjHi20fUuVKW49Z7lvStHcvOHK
c4XyP+e3S/wg6VEMT64kAuZUvvRLhUvJK9ZfxlEGZnjQ+qrYCEpGGDjeDTeXOxMi
6NL7Rh09p/yjemw3u+EZkfNlBMgBsA2+zEOxKbAGmENjjctFGRTJVGKq+FWR2HMi
P2pHCOPEcn2on3GAyTncdp1ANcBNTjb8gTnsoPbc
-----END CERTIFICATE-----
OPENSHIFT_DEPLOYMENT_NAME: docker-registry-2
OPENSHIFT_DEPLOYMENT_NAMESPACE: default
Conditions:
Type Status
PodScheduled False
Volumes:
deployer-token-sbvm4:
Type: Secret (a volume populated by a Secret)
SecretName: deployer-token-sbvm4
QoS Class: BestEffort
Tolerations: <none>
Events:
FirstSeen LastSeen Count From
SubobjectPath Type Reason Message
--------- -------- ----- ----
------------- -------- ------ -------
11m 7m 4 {default-scheduler }
Warning FailedScheduling pod (docker-registry-2-deploy)
failed to fit in any node
fit failure on node (node2-home): CheckServiceAffinity, MatchNodeSelector
fit failure on node (node1-home): MatchNodeSelector, CheckServiceAffinity
4m 3m 2 {default-scheduler } Warning
FailedScheduling pod (docker-registry-2-deploy) failed to fit in any
node
fit failure on node (node2-home): CheckServiceAffinity, MatchNodeSelector
fit failure on node (node1-home): CheckServiceAffinity, MatchNodeSelector
11m 2m 3 {default-scheduler } Warning
FailedScheduling pod (docker-registry-2-deploy) failed to fit in any
node
fit failure on node (node2-home): MatchNodeSelector, CheckServiceAffinity
fit failure on node (node1-home): CheckServiceAffinity, MatchNodeSelector
13m 2m 13 {default-scheduler } Warning
FailedScheduling pod (docker-registry-2-deploy) failed to fit in any
node
fit failure on node (node1-home): MatchNodeSelector, CheckServiceAffinity
fit failure on node (node2-home): MatchNodeSelector, CheckServiceAffinity
11m 1m 7 {default-scheduler } Warning
FailedScheduling pod (docker-registry-2-deploy) failed to fit in any
node
fit failure on node (node1-home): CheckServiceAffinity, MatchNodeSelector
fit failure on node (node2-home): MatchNodeSelector, CheckServiceAffinity
13m 1m 4 {default-scheduler } Warning
FailedScheduling pod (docker-registry-2-deploy) failed to fit in any
node
fit failure on node (node2-home): MatchNodeSelector, CheckServiceAffinity
fit failure on node (node1-home): MatchNodeSelector, CheckServiceAffinity
13m 37s 10 {default-scheduler } Warning
FailedScheduling pod (docker-registry-2-deploy) failed to fit in any
node
fit failure on node (node1-home): MatchNodeSelector, CheckServiceAffinity
fit failure on node (node2-home): CheckServiceAffinity, MatchNodeSelector
13m 5s 7 {default-scheduler } Warning
FailedScheduling pod (docker-registry-2-deploy) failed to fit in any
node
fit failure on node (node1-home): CheckServiceAffinity, MatchNodeSelector
fit failure on node (node2-home): CheckServiceAffinity, MatchNodeSelector
# oc describe pod router-2-deploy
Name: router-2-deploy
Namespace: default
Security Policy: restricted
Node: /
Labels: openshift.io/deployer-pod-for.name=router-2
Status: Pending
IP:
Controllers: <none>
Containers:
deployment:
Image: openshift/origin-deployer:v1.4.1
Port:
Volume Mounts:
/var/run/secrets/kubernetes.io/serviceaccount from
deployer-token-sbvm4 (ro)
Environment Variables:
KUBERNETES_MASTER: https://master-home:8443
OPENSHIFT_MASTER: https://master-home:8443
BEARER_TOKEN_FILE: /var/run/secrets/
kubernetes.io/serviceaccount/token
OPENSHIFT_CA_DATA: -----BEGIN CERTIFICATE-----
MIIC6jCCAdKgAwIBAgIBATANBgkqhkiG9w0BAQsFADAmMSQwIgYDVQQDDBtvcGVu
c2hpZnQtc2lnbmVyQDE0OTU1NDE2MTEwHhcNMTcwNTIzMTIxMzMwWhcNMjIwNTIy
MTIxMzMxWjAmMSQwIgYDVQQDDBtvcGVuc2hpZnQtc2lnbmVyQDE0OTU1NDE2MTEw
ggEiMA0GCSqGSIb3DQEBAQUAA4IBDwAwggEKAoIBAQC9imtjAe8JBjpD99nt3D4h
VCwlWKCMugpIGWYdnHaICBS71KuIim8pWaOWYPUb73QhoUUZhZ80MYOzlB7lk/xK
NWUnQBDFYc9zKqXkxjiWlTXHv1UCyB56mxFdfxPTHN61JbE8dD9jbiBLRudgb1cq
Vhff4CRXqkdDURk8KjpnGkWW57Ky0Icp0rbOrRT/OhYv5CB8sqJedSC2VKfe9qtz
+L4ykOOa4Q1qfqD7YqPDAqnUEJFXEbqjFCdLe6q2TS0vscx/rRJcANmzApgw4BRd
OxEHH1KX6ariXSNkSWxhQIBa8qDukDrGc2dvAoLHi8ALBbnpGLE0zwtf087zdyF/
AgMBAAGjIzAhMA4GA1UdDwEB/wQEAwICpDAPBgNVHRMBAf8EBTADAQH/MA0GCSqG
SIb3DQEBCwUAA4IBAQA7Nn3iGUVH0HJN6WxR6oirpIv9VdqRgugqOoBM8O5GlV7D
7kd4VGFSzFtXKr0kHCgA+/6sEiu0ZlQZT7IvwDWgiY/bXOY/gT8whMWVLXbXBGGT
4brdqSRQVdgjv56kBG/cqIWedwNzItGFb+eye+AjHi20fUuVKW49Z7lvStHcvOHK
c4XyP+e3S/wg6VEMT64kAuZUvvRLhUvJK9ZfxlEGZnjQ+qrYCEpGGDjeDTeXOxMi
6NL7Rh09p/yjemw3u+EZkfNlBMgBsA2+zEOxKbAGmENjjctFGRTJVGKq+FWR2HMi
P2pHCOPEcn2on3GAyTncdp1ANcBNTjb8gTnsoPbc
-----END CERTIFICATE-----
OPENSHIFT_DEPLOYMENT_NAME: router-2
OPENSHIFT_DEPLOYMENT_NAMESPACE: default
Conditions:
Type Status
PodScheduled False
Volumes:
deployer-token-sbvm4:
Type: Secret (a volume populated by a Secret)
SecretName: deployer-token-sbvm4
QoS Class: BestEffort
Tolerations: <none>
Events:
FirstSeen LastSeen Count From
SubobjectPath Type Reason Message
--------- -------- ----- ----
------------- -------- ------ -------
13m 13m 1 {default-scheduler }
Warning FailedScheduling pod (router-2-deploy) failed
to fit in any node
fit failure on node (node2-home): MatchNodeSelector, CheckServiceAffinity
fit failure on node (node1-home): CheckServiceAffinity, MatchNodeSelector
14m 10m 2 {default-scheduler } Warning
FailedScheduling pod (router-2-deploy) failed to fit in any node
fit failure on node (node2-home): CheckServiceAffinity, MatchNodeSelector
fit failure on node (node1-home): MatchNodeSelector, CheckServiceAffinity
15m 5m 12 {default-scheduler } Warning
FailedScheduling pod (router-2-deploy) failed to fit in any node
fit failure on node (node1-home): CheckServiceAffinity, MatchNodeSelector
fit failure on node (node2-home): CheckServiceAffinity, MatchNodeSelector
11m 4m 3 {default-scheduler } Warning
FailedScheduling pod (router-2-deploy) failed to fit in any node
fit failure on node (node2-home): CheckServiceAffinity, MatchNodeSelector
fit failure on node (node1-home): CheckServiceAffinity, MatchNodeSelector
14m 1m 10 {default-scheduler } Warning
FailedScheduling pod (router-2-deploy) failed to fit in any node
fit failure on node (node1-home): CheckServiceAffinity, MatchNodeSelector
fit failure on node (node2-home): MatchNodeSelector, CheckServiceAffinity
15m 54s 12 {default-scheduler } Warning
FailedScheduling pod (router-2-deploy) failed to fit in any node
fit failure on node (node1-home): MatchNodeSelector, CheckServiceAffinity
fit failure on node (node2-home): MatchNodeSelector, CheckServiceAffinity
15m 46s 11 {default-scheduler } Warning
FailedScheduling pod (router-2-deploy) failed to fit in any node
fit failure on node (node1-home): MatchNodeSelector, CheckServiceAffinity
fit failure on node (node2-home): CheckServiceAffinity, MatchNodeSelector
15m 30s 5 {default-scheduler } Warning
FailedScheduling pod (router-2-deploy) failed to fit in any node
fit failure on node (node2-home): MatchNodeSelector, CheckServiceAffinity
fit failure on node (node1-home): MatchNodeSelector, CheckServiceAffinity
I think that there's something wrong with my ansible host file - here it
is, specially the last few lines:
[OSEv3:children]
masters
nodes
[OSEv3:vars]
ansible_ssh_user=root
# ansible_become=yes
deployment_type=origin
openshift_release=v1.4
openshift_image_tag=v1.4.1
containerized=true
openshift_install_examples=true
# openshift_hosted_metrics_deploy=true
# use htpasswd authentication with demo/demo
openshift_master_identity_providers=[{'name': 'htpasswd_auth', 'login':
'true', 'challenge': 'true', 'kind': 'HTPasswdPasswordIdentityProvider',
'filename': '/etc/origin/master/htpasswd'}]
openshift_master_htpasswd_users={'demo':
'$apr1$.MaA77kd$Rlnn6RXq9kCjnEfh5I3w/.'}
# put the router on dedicated infra node
openshift_hosted_router_selector='region=infra'
openshift_master_default_subdomain=apps.hetzlabs.pro
# put the image registry on dedicated infra node
openshift_hosted_registry_selector='region=infra'
# project pods should be placed on primary nodes
osm_default_node_selector='region=primary'
[masters]
master-home.hetzlabs.pro openshift_public_hostname="master-home.hetzlabs.pro
"
[nodes]
# master needs to be included in the node to be configured in the SDN
master-home.hetzlabs.pro openshift_node_labels="{'region': 'infra', 'zone':
'default'}"
node1-home.hetzlabs.pro openshift_node_labels="{'region': 'primary',
'zone': 'default'}"
node2-home.hetzlabs.pro openshift_node_labels="{'region': 'primary',
'zone': 'default'}"
Basically I'm looking to run openshift origin (1.4 or 1.5) with 1 master, 2
nodes (total: 3 vm's). Am I doing it right?
Thanks
On Tue, May 23, 2017 at 4:24 PM, Rodrigo Bersa <[email protected]> wrote:
> Hi Hetz,
>
> It seems that your Registry and Router PODs are not running. Probably
> there's a problem avoiding them to deploy.
>
> Can you send the output of the commands below?
>
> # oc describe pod docker-registry-1-deploy
> # oc describe pod router-1-deploy
>
>
>
> Rodrigo Bersa
>
> Cloud Consultant, RHCSA, RHCVA
>
> Red Hat Brasil <https://www.redhat.com>
>
> [email protected] M: +55 11 99557-5841 <+55-11-99557-5841>
> <https://red.ht/sig> [image: Red Hat] <http://www.redhat.com.br>
> TRIED. TESTED. TRUSTED. <https://redhat.com/trusted>
>
>
> <http://www.redhat.com.br>
>
> On Tue, May 23, 2017 at 8:28 AM, Hetz Ben Hamo <[email protected]> wrote:
>
>> ]# oc get pods -n default
>> NAME READY STATUS RESTARTS AGE
>> docker-registry-1-deploy 0/1 Pending 0 16m
>> registry-console-1-deploy 0/1 Error 0 15m
>> router-1-deploy 0/1 Pending 0 17m
>> [root@master-home ~]# oc logs registry-console-1-deploy
>> --> Scaling registry-console-1 to 1
>> --> Waiting up to 10m0s for pods in rc registry-console-1 to become ready
>> error: update acceptor rejected registry-console-1: pods for rc
>> "registry-console-1" took longer than 600 seconds to become ready
>> [root@master-home ~]# oc logs router-1-deploy
>> [root@master-home ~]# oc logs docker-registry-1-deploy
>> [root@master-home ~]# oc logs docker-registry-1-deploy -n default
>> [root@master-home ~]# oc get pods
>>
>>
>> תודה,
>> *חץ בן חמו*
>> אתם מוזמנים לבקר בבלוג היעוץ <http://linvirtstor.net/> או בבלוג הפרטי שלי
>> <http://benhamo.org>
>>
>> On Tue, May 23, 2017 at 1:49 AM, Ben Parees <[email protected]> wrote:
>>
>>>
>>>
>>> On Mon, May 22, 2017 at 6:18 PM, Hetz Ben Hamo <[email protected]> wrote:
>>>
>>>> Hi,
>>>>
>>>> I've built on a 3 nodes openshift origin using the host file included
>>>> below, but it seems few things are getting broken. I didn't modify anything
>>>> yet on the openshift, just used the openshift-Ansible checked out from
>>>> today.
>>>>
>>>> Problem one: After building an image from the examples (I chose Java
>>>> with the example of wildfly) I get:
>>>>
>>>> [INFO] ------------------------------------------------------------
>>>> ------------
>>>> [INFO] BUILD SUCCESS
>>>> [INFO] ------------------------------------------------------------
>>>> ------------
>>>> [INFO] Total time: 12.182 s
>>>> [INFO] Finished at: 2017-05-22T22:08:21+00:00
>>>> [INFO] Final Memory: 14M/134M
>>>> [INFO] ------------------------------------------------------------
>>>> ------------
>>>> Moving built war files into /wildfly/standalone/deployments for later
>>>> deployment...
>>>> Moving all war artifacts from /opt/app-root/src/target directory into
>>>> /wildfly/standalone/deployments for later deployment...
>>>> '/opt/app-root/src/target/ROOT.war' -> '/wildfly/standalone/deploymen
>>>> ts/ROOT.war'
>>>> Moving all ear artifacts from /opt/app-root/src/target directory into
>>>> /wildfly/standalone/deployments for later deployment...
>>>> Moving all rar artifacts from /opt/app-root/src/target directory into
>>>> /wildfly/standalone/deployments for later deployment...
>>>> Moving all jar artifacts from /opt/app-root/src/target directory into
>>>> /wildfly/standalone/deployments for later deployment...
>>>> ...done
>>>> Pushing image 172.30.172.85:5000/test1/wf:latest ...
>>>> Warning: Push failed, retrying in 5s ...
>>>> Warning: Push failed, retrying in 5s ...
>>>> Warning: Push failed, retrying in 5s ...
>>>> Warning: Push failed, retrying in 5s ...
>>>> Warning: Push failed, retrying in 5s ...
>>>> Warning: Push failed, retrying in 5s ...
>>>> Warning: Push failed, retrying in 5s ...
>>>> Registry server Address:
>>>> Registry server User Name: serviceaccount
>>>> Registry server Email: [email protected]
>>>> Registry server Password: <<non-empty>>
>>>> error: build error: Failed to push image: Get
>>>> https://172.30.172.85:5000/v1/_ping: dial tcp 172.30.172.85:5000:
>>>> getsockopt: connection refused
>>>>
>>>>
>>> can you confirm your registry pod is running in the default namespace
>>> (oc get pods -n default)? Can you get logs from it?
>>>
>>>
>>>
>>>>
>>>> Another problem: I added the metrics option so it installed hawkler but
>>>> when it complains that it needs SSL approval (it shows a message about a
>>>> problem with hawkler and gives a link to open it) I get upon clicking the
>>>> link: connection refused.
>>>>
>>>> I've tested the host configuration on 2 sets of VM's (one at home with
>>>> digital ocean, another set here at home with VMWare). I've set up DNS with
>>>> subdomain wildcard and I can ping the app names but trying to connect
>>>> through a browser or curl - gives connection refused.
>>>>
>>>> Have I missed something?
>>>>
>>>> here is my byo host file:
>>>>
>>>> [OSEv3:children]
>>>> masters
>>>> nodes
>>>>
>>>> [OSEv3:vars]
>>>> ansible_ssh_user=root
>>>>
>>>> deployment_type=origin
>>>> openshift_release=v1.5.0
>>>> containerized=true
>>>> openshift_install_examples=true
>>>> openshift_hosted_metrics_deploy=true
>>>>
>>>> # use htpasswd authentication with demo/demo
>>>> openshift_master_identity.providers=[{'name': 'htpasswd_auth',
>>>> 'login': 'true', 'challenge': 'true', 'kind':
>>>> 'HTPasswdPasswordIdentityProvider',
>>>> 'filename': '/etc/origin/master/htpasswd'}]
>>>> openshift_master_htpasswd_users={'demo': '$XXXXXXXXXXXXXXXXXXXXXXXXXXXX
>>>> XXXXXXXX.'}
>>>>
>>>> # put the router on dedicated infra node
>>>> openshift_hosted_router_selector='region=infra'
>>>> openshift_master_default_subdomain=apps.test.com
>>>>
>>>> # put the image registry on dedicated infra node
>>>> openshift_hosted_registry_selector='region=infra'
>>>>
>>>> #.project pods should be placed on primary nodes
>>>> osm_default_node_selector='region=primary'
>>>>
>>>> [masters]
>>>> master-home.test.com openshift_public_hostname="master-home.test.com"
>>>>
>>>> [nodes]
>>>> # master needs to be included in the node to be configured in the SDN
>>>> # master-home.test.com
>>>> master-home.test.com openshift_node_labels="{'region': 'infra',
>>>> 'zone': 'default'}"
>>>> node1-home.test.com openshift_node_labels="{'region': 'primary',
>>>> 'zone': 'default'}"
>>>> node2-home.test.com openshift_node_labels="{'region': 'primary',
>>>> 'zone': 'default'}"
>>>>
>>>>
>>>> Thanks
>>>>
>>>> _______________________________________________
>>>> users mailing list
>>>> [email protected]
>>>> http://lists.openshift.redhat.com/openshiftmm/listinfo/users
>>>>
>>>>
>>>
>>>
>>> --
>>> Ben Parees | OpenShift
>>>
>>>
>>
>> _______________________________________________
>> users mailing list
>> [email protected]
>> http://lists.openshift.redhat.com/openshiftmm/listinfo/users
>>
>>
>
_______________________________________________
users mailing list
[email protected]
http://lists.openshift.redhat.com/openshiftmm/listinfo/users