I've managed to proceed with my deployment. I've changed my /etc/ansible/hosts 
file and my deployment was successful.

In ca se someone encounters similar issues, below it is the /etc/ansible/hosts 
file I have used:

[OSEv3:children]
masters
etcd
nodes

[OSEv3:vars]
openshift_master_default_subdomain=apps.mydomain.com
ansible_ssh_user=root
openshift_release=v3.7.0
openshift_image_tag=v3.7.0
openshift_pkg_version=-3.7.0
openshift_master_cluster_method=native
openshift_master_cluster_hostname=internal-master-centos.mydomain.com
openshift_master_cluster_public_hostname=master-centos.mydomain.com
openshift_deployment_type=origin
openshift_disable_check=memory_availability,disk_availability,docker_image_availability
containerized=true
openshift_docker_options='--selinux-enabled --insecure-registry 172.30.0.0/16'

openshift_master_identity_providers=[{'name': 'htpasswd_auth', 'login': 'true', 
'challenge': 'true', 'kind': 'HTPasswdPasswordIdentityProvider', 'filename': 
'/etc/origin/htpasswd'}]

[masters]
<Master IP Address>

[etcd]
<Master IP Address>

[nodes]
<Master IP Address>
<Slave IP Address> openshift_node_labels="{'region':'infra','zone':'default'}"


Thanks,
Anda
From: users-boun...@lists.openshift.redhat.com 
[mailto:users-boun...@lists.openshift.redhat.com] On Behalf Of Anda Nicolae
Sent: Friday, March 2, 2018 12:22 PM
To: users@lists.openshift.redhat.com
Subject: Error response from daemon: No such container: origin-node when 
Deploying OpenShift Origin v3.7.0 on CentOS 7

Hi all,

I have a CentOS 7 server and I have created 2 virtual machines (VMS) on this 
server. CentOS 7 is also the OS that runs on the VMs.
I am trying to deploy OpenShift Origin v3.7.0 on these 2 VMs (1st VM is the 
master and the 2nd VM is the node)

I have run the following commands:
ansible-playbook -i /etc/ansible/hosts 
/root/openshift-ansible/playbooks/prerequisites.yml
which successfully finishes, and:

ansible-playbook -vvv -i /etc/ansible/hosts 
/root/openshift-ansible/playbooks/deploy_cluster.yml
which fails with the following error:

fatal: [<Master IP Address>]: FAILED! => {
    "attempts": 3,
    "changed": false,
    "invocation": {
        "module_args": {
            "daemon_reload": false,
            "enabled": null,
            "masked": null,
            "name": "origin-node",
            "no_block": false,
            "state": "restarted",
            "user": false
        }
    }
}

MSG:

Unable to restart service origin-node: Job for origin-node.service failed 
because the control process exited with error code. See "systemctl status 
origin-node.service" and "journalctl -xe" for de                                
                           tails.


META: ran handlers
        to retry, use: --limit 
@/root/openshift-ansible/playbooks/deploy_cluster.retry

PLAY RECAP 
***************************************************************************************
<Master IP Address>              : ok=407  changed=75   unreachable=0    
failed=1
<Node IP Address>                 : ok=53   changed=6    unreachable=0    
failed=0
localhost                                   : ok=13   changed=0    
unreachable=0    failed=0

>From journalctl -xe, I have:
umounthook <error>: c0be9a269c90: Failed to read directory 
/usr/share/oci-umount/oci-umount.d: No such file or directory
master-centos.mydomain.com origin-node[49477]: Error response from daemon: No 
such container: origin-node

docker ps shows:
[root@master-centos ~]# docker ps
CONTAINER ID        IMAGE                                    COMMAND            
      CREATED             STATUS              PORTS               NAMES
6e36d7f538da        openshift/origin:v3.7.1                  
"/usr/bin/openshift s"   29 minutes ago      Up 29 minutes                      
     origin-master-controllers
2c4136affcd5        openshift/origin:v3.7.1                  
"/usr/bin/openshift s"   29 minutes ago      Up 29 minutes                      
     origin-master-api
308a746e77f5        registry.fedoraproject.org/latest/etcd   "/usr/bin/etcd"    
      29 minutes ago      Up 29 minutes                           etcd_container
ef9f2c58c8f7        openshift/openvswitch:v3.7.1             
"/usr/local/bin/ovs-r"   29 minutes ago      Up 29 minutes


Below is my hosts file:
cat /etc/ansible/hosts
masters
etcd
nodes

[OSEv3:vars]
openshift_master_default_subdomain=apps.mydomain.com
ansible_ssh_user=root
ansible_become=yes
containerized=true
openshift_release=v3.7
openshift_image_tag=v3.7.1
openshift_pkg_version=-3.7.0
openshift_master_cluster_method=native
openshift_master_cluster_hostname=internal-master-centos.mydomain.com
openshift_master_cluster_public_hostname=master-centos.mydomain.com
deployment_type=origin

os_sdn_network_plugin_name='redhat/openshift-ovs-multitenant'
openshift_master_overwrite_named_certificates=true
openshift_master_identity_providers=[{'name': 'htpasswd_auth', 'login': 'true', 
'challenge': 'true', 'kind': 'HTPasswdPasswordIdentityProvider', 'filename': 
'/etc/origin/htpasswd'}]
openshift_docker_options='--selinux-enabled --insecure-registry 172.30.0.0/16'
openshift_router_selector='region=infra'
openshift_registry_selector='region=infra'
openshift_master_api_port=443
openshift_master_console_port=443
openshift_disable_check=memory_availability,disk_availability,docker_image_availability

[masters]
<Master IP Address> openshift_hostname=master-centos.mydomain.com

[etcd]
<Master IP Address> openshift_hostname=master-centos.mydomain.com

[nodes]
<Master IP Address> openshift_hostname=master-centos.mydomain.com 
openshift_schedulable=false
<Node IP Address> openshift_hostname=slave-centos.mydomain.com 
openshift_node_labels="{'router':'true','registry':'true'}"

Do you have any idea why my deployment fails?

Thanks,
Anda
_______________________________________________
users mailing list
users@lists.openshift.redhat.com
http://lists.openshift.redhat.com/openshiftmm/listinfo/users

Reply via email to