Re: questions about externalIP usage

2017-08-15 Thread Yu Wei
Hi Erik,

I configured "externalIPs" with cluster setup using flannel but it didn't work.

Then I reinstalled cluster with openvswitch. This time "externalIP" worked.


Is openvwitch suggested network plugin for openshift origin?

I deployed same application in k8s cluster and it worked.


Thanks,

Jared, (韦煜)
Software developer
Interested in open source software, big data, Linux


From: Erik Jacobs <ejac...@redhat.com>
Sent: Wednesday, August 16, 2017 8:16:07 AM
To: Yu Wei
Cc: us...@lists.openshift.redhat.com; dev@lists.openshift.redhat.com
Subject: Re: questions about externalIP usage

Hi Jared,

Did you previously configure the cluster for externalip usage?

https://docs.openshift.org/latest/admin_guide/tcp_ingress_external_ports.html

---

ERIK JACOBS

PRINCIPAL TECHNICAL MARKETING MANAGER, OPENSHIFT

Red Hat Inc<https://www.redhat.com/>

ejac...@redhat.com<mailto:ejac...@redhat.com>M: 
646.462.3745 @: erikonopen

[https://www.redhat.com/files/brand/email/sig-redhat.png]<https://red.ht/sig>
TRIED. TESTED. TRUSTED.<https://redhat.com/trusted>


On Thu, Aug 10, 2017 at 4:12 AM, Yu Wei 
<yu20...@hotmail.com<mailto:yu20...@hotmail.com>> wrote:

Hi guys,

I deployed redis with replication controller successfully on openshift origin 
cluster.

Then I tried to create service for external clients to connect.

However, it seemed that it didn't work.

How could I debug similar problem? Is there any guidance about using externalIP 
in openshift?


The detailed information is as below,

[root@host-10-1-236-92 gluster]# oc get svc
NAMECLUSTER-IP EXTERNAL-IP   
PORT(S)  AGE
glusterfs-cluster   172.30.6.143   1/TCP  
  1d
redis-svc   172.30.51.20   10.1.236.92,10.1.236.93,10.1.241.55   
26379/TCP,6379/TCP   24m
[root@host-10-1-236-92 gluster]# oc describe svc redis-svc
Name:redis-svc
Namespace:openshiift-servicebroker
Labels:
Selector:sb-2017-redis-master=master
Type:ClusterIP
IP:172.30.51.20
Port:redis-sen26379/TCP
Endpoints:172.30.41.5:26379<http://172.30.41.5:26379>
Port:redis-master6379/TCP
Endpoints:172.30.41.5:6379<http://172.30.41.5:6379>
Session Affinity:None
No events.
[root@host-10-1-236-92 gluster]# cat redis-master-svc.yaml
---
kind: Service
apiVersion: v1
metadata:
  name: redis-svc
spec:
selector:
  sb-2017-redis-master: master
ports:
  - name: redis-sen
protocol: TCP
port: 26379
targetPort: 26379
  - name: redis-master
protocol: TCP
port: 6379
targetPort: 6379
externalIPs:
  -  10.1.236.92
  -  10.1.236.93
  -  10.1.241.55



Thanks,

Jared, (韦煜)
Software developer
Interested in open source software, big data, Linux

___
dev mailing list
dev@lists.openshift.redhat.com<mailto:dev@lists.openshift.redhat.com>
http://lists.openshift.redhat.com/openshiftmm/listinfo/dev


___
dev mailing list
dev@lists.openshift.redhat.com
http://lists.openshift.redhat.com/openshiftmm/listinfo/dev


Re: HAProxy not running on master

2017-08-15 Thread Yu Wei
Yes, I set master as work nodes and as infra nodes.

Thanks for your help. I understood this concept now.



Thanks,

Jared, (韦煜)
Software developer
Interested in open source software, big data, Linux


From: Akram Ben Aissi <akram.benai...@gmail.com>
Sent: Tuesday, August 15, 2017 7:11:03 PM
To: Yu Wei
Cc: us...@lists.openshift.redhat.com; dev@lists.openshift.redhat.com
Subject: Re: HAProxy not running on master

Hi Yui,

the masters runs the OpenShift API and web console on port 8443. OpenShift CLI 
or nodes communicates with master on this port.
OpenShift router are not really supposed to run on masters, unless you set your 
masters also as nodes, and specifically as infra nodes (region=infra)

If you do so, then, indeed, you will have an additional HAProxy process running 
on masters and listening on ports 80 and 443.




On 15 August 2017 at 12:47, Yu Wei 
<yu20...@hotmail.com<mailto:yu20...@hotmail.com>> wrote:

Hi,


I setup cluster with 1 lb, 2 masters and 2 work nodes.

On host lb, haproxy service and openshift_loadbalancer was running.

On master node, I found that haproxy was failed to start. And 
openshift/origin-haproxy-router was running with container name 
k8s_router.xxx.

I also observed that on that master node many connections was established via 
port 8443.


Is this expected behavior? Or anything goes wrong?


Thanks,

Jared, (韦煜)
Software developer
Interested in open source software, big data, Linux

___
dev mailing list
dev@lists.openshift.redhat.com<mailto:dev@lists.openshift.redhat.com>
http://lists.openshift.redhat.com/openshiftmm/listinfo/dev


___
dev mailing list
dev@lists.openshift.redhat.com
http://lists.openshift.redhat.com/openshiftmm/listinfo/dev


Different behavior between installing openshift origin with openvswitch and flannel

2017-08-15 Thread Yu Wei
Hi guys,

I tried to get external traffic into openshift origin cluster using 
nodeport/externalIPs.

When I setup openshift cluster with flannel, exposing service with 
nodeport/externalIPs did not work.

When switched to openvswitch, both worked.


Is this expected behavior? Or did I miss anything?


Thanks,

Jared, (韦煜)
Software developer
Interested in open source software, big data, Linux
___
dev mailing list
dev@lists.openshift.redhat.com
http://lists.openshift.redhat.com/openshiftmm/listinfo/dev


HAProxy not running on master

2017-08-15 Thread Yu Wei
Hi,


I setup cluster with 1 lb, 2 masters and 2 work nodes.

On host lb, haproxy service and openshift_loadbalancer was running.

On master node, I found that haproxy was failed to start. And 
openshift/origin-haproxy-router was running with container name 
k8s_router.xxx.

I also observed that on that master node many connections was established via 
port 8443.


Is this expected behavior? Or anything goes wrong?


Thanks,

Jared, (韦煜)
Software developer
Interested in open source software, big data, Linux
___
dev mailing list
dev@lists.openshift.redhat.com
http://lists.openshift.redhat.com/openshiftmm/listinfo/dev


questions about externalIP usage

2017-08-10 Thread Yu Wei
Hi guys,

I deployed redis with replication controller successfully on openshift origin 
cluster.

Then I tried to create service for external clients to connect.

However, it seemed that it didn't work.

How could I debug similar problem? Is there any guidance about using externalIP 
in openshift?


The detailed information is as below,

[root@host-10-1-236-92 gluster]# oc get svc
NAMECLUSTER-IP EXTERNAL-IP   
PORT(S)  AGE
glusterfs-cluster   172.30.6.143   1/TCP  
  1d
redis-svc   172.30.51.20   10.1.236.92,10.1.236.93,10.1.241.55   
26379/TCP,6379/TCP   24m
[root@host-10-1-236-92 gluster]# oc describe svc redis-svc
Name:redis-svc
Namespace:openshiift-servicebroker
Labels:
Selector:sb-2017-redis-master=master
Type:ClusterIP
IP:172.30.51.20
Port:redis-sen26379/TCP
Endpoints:172.30.41.5:26379
Port:redis-master6379/TCP
Endpoints:172.30.41.5:6379
Session Affinity:None
No events.
[root@host-10-1-236-92 gluster]# cat redis-master-svc.yaml
---
kind: Service
apiVersion: v1
metadata:
  name: redis-svc
spec:
selector:
  sb-2017-redis-master: master
ports:
  - name: redis-sen
protocol: TCP
port: 26379
targetPort: 26379
  - name: redis-master
protocol: TCP
port: 6379
targetPort: 6379
externalIPs:
  -  10.1.236.92
  -  10.1.236.93
  -  10.1.241.55



Thanks,

Jared, (韦煜)
Software developer
Interested in open source software, big data, Linux
___
dev mailing list
dev@lists.openshift.redhat.com
http://lists.openshift.redhat.com/openshiftmm/listinfo/dev


Etcd issue during openshift origin installation

2017-07-21 Thread Yu Wei
Hi guys,

I tried to setup openshift origin cluster using "Advanced Installation".

In file hosts, I added following variables,

containerized=true

osm_etcd_image=registry.access.redhat.com/rhel7/etcd

Per my understanding, etcd will run in docker container.
However, during installation, I got following error?
TASK [etcd_common : Install etcd for etcdctl] 
*
fatal: [host-10-1-236-92]: FAILED! => {"changed": false, "failed": true, "msg": 
"No package matching 'etcd' found available, installed or updated", "rc": 126, 
"results": ["No package matching 'etcd' found available, installed or updated"]}

After installing etcd RPM packages manually on nodes, installation process 
continued.


What's the functionality of ectd here? Is it running in docker container?


Thanks,

Jared, (韦煜)
Software developer
Interested in open source software, big data, Linux
___
dev mailing list
dev@lists.openshift.redhat.com
http://lists.openshift.redhat.com/openshiftmm/listinfo/dev


Re: Docker related issues when installing openshift origin 1.5 on Redhat 7.2

2017-07-21 Thread Yu Wei
Thanks for the information.

I missed this section when I read the documentation.


Thanks,

Jared, (韦煜)
Software developer
Interested in open source software, big data, Linux


From: Scott Dodson <sdod...@redhat.com>
Sent: Friday, July 21, 2017 9:52:57 PM
To: Yu Wei
Cc: Łukasz Strzelec; us...@lists.openshift.redhat.com; 
dev@lists.openshift.redhat.com
Subject: Re: Docker related issues when installing openshift origin 1.5 on 
Redhat 7.2

For RHEL this section of the documentation should cover all of the registration 
and repo requirements
https://docs.openshift.com/container-platform/3.5/install_config/install/host_preparation.html#host-registration

For Centos I think this will work, but the repo should be enabled by default 
already.
yum-config-manager --enable extras

On Fri, Jul 21, 2017 at 9:28 AM, Yu Wei 
<yu20...@hotmail.com<mailto:yu20...@hotmail.com>> wrote:

Hi Scott,

Yes, repo used for installing docker is from 
yum.dockerproject.com<http://yum.dockerproject.com>.

Another basic question is how could I enable RHEL docker repo to install docker.

I'm not familiar with redhat.


Thanks,

Jared, (韦煜)
Software developer
Interested in open source software, big data, Linux


From: Scott Dodson <sdod...@redhat.com<mailto:sdod...@redhat.com>>
Sent: Friday, July 21, 2017 9:07:01 PM
To: Łukasz Strzelec
Cc: Yu Wei; 
us...@lists.openshift.redhat.com<mailto:us...@lists.openshift.redhat.com>; 
dev@lists.openshift.redhat.com<mailto:dev@lists.openshift.redhat.com>
Subject: Re: Docker related issues when installing openshift origin 1.5 on 
Redhat 7.2

You have docker from the upstream repos installed and unfortunately we're not 
compatible with their packaging. Please use docker from rhel/centos/fedora 
repos. You may have to enable the extras repo if it's not currently enabled and 
remove all docker packages before retrying.


--
Scott

On Fri, Jul 21, 2017 at 6:40 AM, Łukasz Strzelec 
<lukasz.strze...@gmail.com<mailto:lukasz.strze...@gmail.com>> wrote:
Hello :)
Can you share with us your inventory file?
Openshift provides several additional variables realted to your issue, see 
below:

# Specify exact version of Docker to configure or upgrade to.
# Downgrades are not supported and will error out. Be careful when upgrading 
docker from < 1.10 to > 1.10.
# docker_version="1.12.1"

# Skip upgrading Docker during an OpenShift upgrade, leaves the current Docker 
version alone.
# docker_upgrade=False

Best regards

2017-07-19 11:52 GMT+02:00 Yu Wei 
<yu20...@hotmail.com<mailto:yu20...@hotmail.com>>:

Hi guys,

I tried to install openshift origin 1.5 on Redhat with docker 1.12.6 installed 
on each host.

However, it seeded that ansible tried to install docker 1.9 instead of using 
1.12.6.

Is this expected behavior? If not, how could I fix such problem?

The detailed error is as below,


TASK [docker : Error out if attempting to upgrade Docker across the 1.10 
boundary] 
skipping: [host-10-1-236-92]

TASK [docker : Install Docker] 

fatal: [host-10-1-236-92]: FAILED! => {"changed": true, "failed": true, "msg": 
"Error: docker-engine conflicts with 
docker-1.9.1-25.1.origin.el7.x86_64\nError: docker-engine-selinux conflicts 
with docker-selinux-1.9.1-25.1.origin.el7.x86_64\n", "rc": 1, "results": 
["Loaded plugins: langpacks, product-id, search-disabled-repos, subscription-\n 
 : manager\nThis system is not registered to Red Hat Subscription 
Management. You can use subscription-manager to register.\nResolving 
Dependencies\n--> Running transaction check\n---> Package docker.x86_64 
0:1.9.1-25.1.origin.el7 will be installed\n--> Processing Dependency: 
docker-forward-journald = 1.9.1-25.1.origin.el7 for package: 
docker-1.9.1-25.1.origin.el7.x86_64\n--> Processing Dependency: docker-selinux 
>= 1.9.1-25.1.origin.el7 for package: docker-1.9.1-25.1.origin.el7.x86_64\n--> 
Running transaction check\n---> Package docker-forward-journald.x86_64 
0:1.9.1-25.1.origin.el7 will be installed\n---> Package docker-selinux.x86_64 
0:1.9.1-25.1.origin.el7 will be installed\n--> Processing Conflict: 
docker-engine-1.12.6-1.el7.centos.x86_64 conflicts docker\n--> Processing 
Conflict: docker-engine-1.12.6-1.el7.centos.x86_64 conflicts docker-io\n--> 
Processing Conflict: docker-engine-selinux-1.12.6-1.el7.centos.noarch conflicts 
docker-selinux\n--> Finished Dependency Resolution\n You could try using 
--skip-broken to work around the problem\n You could try running: rpm -Va 
--nofiles --nodigest\n"]}
to retry, use: --limit 
@/root/openshift/openshift-ansible/playbooks/byo/config.r

Re: Docker related issues when installing openshift origin 1.5 on Redhat 7.2

2017-07-21 Thread Yu Wei
Hi Scott,

Yes, repo used for installing docker is from yum.dockerproject.com.

Another basic question is how could I enable RHEL docker repo to install docker.

I'm not familiar with redhat.


Thanks,

Jared, (韦煜)
Software developer
Interested in open source software, big data, Linux


From: Scott Dodson <sdod...@redhat.com>
Sent: Friday, July 21, 2017 9:07:01 PM
To: Łukasz Strzelec
Cc: Yu Wei; us...@lists.openshift.redhat.com; dev@lists.openshift.redhat.com
Subject: Re: Docker related issues when installing openshift origin 1.5 on 
Redhat 7.2

You have docker from the upstream repos installed and unfortunately we're not 
compatible with their packaging. Please use docker from rhel/centos/fedora 
repos. You may have to enable the extras repo if it's not currently enabled and 
remove all docker packages before retrying.


--
Scott

On Fri, Jul 21, 2017 at 6:40 AM, Łukasz Strzelec 
<lukasz.strze...@gmail.com<mailto:lukasz.strze...@gmail.com>> wrote:
Hello :)
Can you share with us your inventory file?
Openshift provides several additional variables realted to your issue, see 
below:

# Specify exact version of Docker to configure or upgrade to.
# Downgrades are not supported and will error out. Be careful when upgrading 
docker from < 1.10 to > 1.10.
# docker_version="1.12.1"

# Skip upgrading Docker during an OpenShift upgrade, leaves the current Docker 
version alone.
# docker_upgrade=False

Best regards

2017-07-19 11:52 GMT+02:00 Yu Wei 
<yu20...@hotmail.com<mailto:yu20...@hotmail.com>>:

Hi guys,

I tried to install openshift origin 1.5 on Redhat with docker 1.12.6 installed 
on each host.

However, it seeded that ansible tried to install docker 1.9 instead of using 
1.12.6.

Is this expected behavior? If not, how could I fix such problem?

The detailed error is as below,


TASK [docker : Error out if attempting to upgrade Docker across the 1.10 
boundary] 
skipping: [host-10-1-236-92]

TASK [docker : Install Docker] 

fatal: [host-10-1-236-92]: FAILED! => {"changed": true, "failed": true, "msg": 
"Error: docker-engine conflicts with 
docker-1.9.1-25.1.origin.el7.x86_64\nError: docker-engine-selinux conflicts 
with docker-selinux-1.9.1-25.1.origin.el7.x86_64\n", "rc": 1, "results": 
["Loaded plugins: langpacks, product-id, search-disabled-repos, subscription-\n 
 : manager\nThis system is not registered to Red Hat Subscription 
Management. You can use subscription-manager to register.\nResolving 
Dependencies\n--> Running transaction check\n---> Package docker.x86_64 
0:1.9.1-25.1.origin.el7 will be installed\n--> Processing Dependency: 
docker-forward-journald = 1.9.1-25.1.origin.el7 for package: 
docker-1.9.1-25.1.origin.el7.x86_64\n--> Processing Dependency: docker-selinux 
>= 1.9.1-25.1.origin.el7 for package: docker-1.9.1-25.1.origin.el7.x86_64\n--> 
Running transaction check\n---> Package docker-forward-journald.x86_64 
0:1.9.1-25.1.origin.el7 will be installed\n---> Package docker-selinux.x86_64 
0:1.9.1-25.1.origin.el7 will be installed\n--> Processing Conflict: 
docker-engine-1.12.6-1.el7.centos.x86_64 conflicts docker\n--> Processing 
Conflict: docker-engine-1.12.6-1.el7.centos.x86_64 conflicts docker-io\n--> 
Processing Conflict: docker-engine-selinux-1.12.6-1.el7.centos.noarch conflicts 
docker-selinux\n--> Finished Dependency Resolution\n You could try using 
--skip-broken to work around the problem\n You could try running: rpm -Va 
--nofiles --nodigest\n"]}
to retry, use: --limit 
@/root/openshift/openshift-ansible/playbooks/byo/config.retry

PLAY RECAP 

host-10-1-236-92   : ok=59   changed=6unreachable=0failed=1




Thanks,

Jared, (韦煜)
Software developer
Interested in open source software, big data, Linux

___
users mailing list
us...@lists.openshift.redhat.com<mailto:us...@lists.openshift.redhat.com>
http://lists.openshift.redhat.com/openshiftmm/listinfo/users




--
Łukasz Strzelec
Sr. DevOps Expert / Product Owner of XaaS platform at ING Services Polska

___
users mailing list
us...@lists.openshift.redhat.com<mailto:us...@lists.openshift.redhat.com>
http://lists.openshift.redhat.com/openshiftmm/listinfo/users


___
dev mailing list
dev@lists.openshift.redhat.com
http://lists.openshift.redhat.com/openshiftmm/listinfo/dev


How could I deploy redis cluster on openshift origin cluster?

2017-07-23 Thread Yu Wei
Hi,

I want to deploy redis cluster on openshfit origin cluster.

Is there any images, deployment that could be used?


Thanks,

Jared, (韦煜)
Software developer
Interested in open source software, big data, Linux
___
dev mailing list
dev@lists.openshift.redhat.com
http://lists.openshift.redhat.com/openshiftmm/listinfo/dev


Re: How could I deploy redis cluster on openshift origin cluster?

2017-07-24 Thread Yu Wei
Thanks for the information.

I want to deploy redis cluster rather than single instance.


Thanks,

Jared, (韦煜)
Software developer
Interested in open source software, big data, Linux


From: Graham Dumpleton <gdump...@redhat.com>
Sent: Monday, July 24, 2017 11:36:08 AM
To: Yu Wei
Cc: us...@lists.openshift.redhat.com; dev@lists.openshift.redhat.com
Subject: Re: How could I deploy redis cluster on openshift origin cluster?

See:

https://github.com/sclorg/redis-container

The image can be found at:

https://hub.docker.com/r/centos/redis-32-centos7/

Graham


On 24 Jul 2017, at 1:26 PM, Yu Wei 
<yu20...@hotmail.com<mailto:yu20...@hotmail.com>> wrote:

Hi,
I want to deploy redis cluster on openshfit origin cluster.
Is there any images, deployment that could be used?

Thanks,
Jared, (韦煜)
Software developer
Interested in open source software, big data, Linux
___
users mailing list
us...@lists.openshift.redhat.com<mailto:us...@lists.openshift.redhat.com>
http://lists.openshift.redhat.com/openshiftmm/listinfo/users

___
dev mailing list
dev@lists.openshift.redhat.com
http://lists.openshift.redhat.com/openshiftmm/listinfo/dev


Docker related issues when installing openshift origin 1.5 on Redhat 7.2

2017-07-19 Thread Yu Wei
Hi guys,

I tried to install openshift origin 1.5 on Redhat with docker 1.12.6 installed 
on each host.

However, it seeded that ansible tried to install docker 1.9 instead of using 
1.12.6.

Is this expected behavior? If not, how could I fix such problem?

The detailed error is as below,


TASK [docker : Error out if attempting to upgrade Docker across the 1.10 
boundary] 
skipping: [host-10-1-236-92]

TASK [docker : Install Docker] 

fatal: [host-10-1-236-92]: FAILED! => {"changed": true, "failed": true, "msg": 
"Error: docker-engine conflicts with 
docker-1.9.1-25.1.origin.el7.x86_64\nError: docker-engine-selinux conflicts 
with docker-selinux-1.9.1-25.1.origin.el7.x86_64\n", "rc": 1, "results": 
["Loaded plugins: langpacks, product-id, search-disabled-repos, subscription-\n 
 : manager\nThis system is not registered to Red Hat Subscription 
Management. You can use subscription-manager to register.\nResolving 
Dependencies\n--> Running transaction check\n---> Package docker.x86_64 
0:1.9.1-25.1.origin.el7 will be installed\n--> Processing Dependency: 
docker-forward-journald = 1.9.1-25.1.origin.el7 for package: 
docker-1.9.1-25.1.origin.el7.x86_64\n--> Processing Dependency: docker-selinux 
>= 1.9.1-25.1.origin.el7 for package: docker-1.9.1-25.1.origin.el7.x86_64\n--> 
Running transaction check\n---> Package docker-forward-journald.x86_64 
0:1.9.1-25.1.origin.el7 will be installed\n---> Package docker-selinux.x86_64 
0:1.9.1-25.1.origin.el7 will be installed\n--> Processing Conflict: 
docker-engine-1.12.6-1.el7.centos.x86_64 conflicts docker\n--> Processing 
Conflict: docker-engine-1.12.6-1.el7.centos.x86_64 conflicts docker-io\n--> 
Processing Conflict: docker-engine-selinux-1.12.6-1.el7.centos.noarch conflicts 
docker-selinux\n--> Finished Dependency Resolution\n You could try using 
--skip-broken to work around the problem\n You could try running: rpm -Va 
--nofiles --nodigest\n"]}
to retry, use: --limit 
@/root/openshift/openshift-ansible/playbooks/byo/config.retry

PLAY RECAP 

host-10-1-236-92   : ok=59   changed=6unreachable=0failed=1




Thanks,

Jared, (韦煜)
Software developer
Interested in open source software, big data, Linux
___
dev mailing list
dev@lists.openshift.redhat.com
http://lists.openshift.redhat.com/openshiftmm/listinfo/dev


openshift origin 1.2 doesn't compatible with docker 1.9

2017-07-22 Thread Yu Wei
Hi guys,

I tried to setup openshift origin cluster 1.2 with docker 1.9.

However, it seemed they're not compatible.

The failure is as below,

RUNNING HANDLER [openshift_node : restart node] 
***
fatal: [host-10-1-236-93]: FAILED! => {"changed": false, "failed": true, "msg": 
"Unable to restart service origin-node: Job for origin-node.service failed 
because the control process exited with error code. See \"systemctl status 
origin-node.service\" and \"journalctl -xe\" for details.\n"}


The root cause is that docker 1.9 doesn't support volume mode "rslave". 
However, node service configuration as follows,

ExecStart=/usr/bin/docker run --name origin-node --rm --privileged --net=host 
--pid=host --env-file=/etc/sysconfig/origin-node -v /:/rootfs:ro,rslave -e 
CONFIG_FILE=${CONFIG_FILE} -e OPTIONS=${OPTIONS} -e HOST=/rootfs -e 
HOST_ETC=/host-etc -v /var/lib/origin:/var/lib/origin -v 
/etc/origin/node:/etc/origin/node -v /etc/localtime:/etc/localtime:ro -v 
/etc/machine-id:/etc/machine-id:ro -v /run:/run -v /sys:/sys:rw -v 
/sys/fs/cgroup:/sys/fs/cgroup:rw -v /usr/bin/docker:/usr/bin/docker:ro -v 
/var/lib/docker:/var/lib/docker -v /lib/modules:/lib/modules -v 
/etc/origin/openvswitch:/etc/openvswitch -v /etc/origin/sdn:/etc/openshift-sdn 
-v /var/lib/cni:/var/lib/cni -v /etc/systemd/system:/host-etc/systemd/system -v 
/var/log:/var/log -v /dev:/dev $DOCKER_ADDTL_BIND_MOUNTS -v 
/etc/pki:/etc/pki:ro openshift/node:${IMAGE_VERSION}


Any advice to workaround this problem? Or is this real issue need to be fixed?


Thanks,

Jared, (韦煜)
Software developer
Interested in open source software, big data, Linux
___
dev mailing list
dev@lists.openshift.redhat.com
http://lists.openshift.redhat.com/openshiftmm/listinfo/dev


Question about router usage

2017-08-01 Thread Yu Wei
Hi guys,

How could I expose services using TCP/UDP protocols to external clients?

Could router be used?

For example, I want to deploy redis cluster in openshift cluster.

Redis cluster is using TCP protocol and listening on port 6379.

Could I expose redis service port 6379 with router?

If not, how could I expose the service to external clients?

Could I use nodePort provided by k8s or other advice?


Thanks,

Jared, (韦煜)
Software developer
Interested in open source software, big data, Linux
___
dev mailing list
dev@lists.openshift.redhat.com
http://lists.openshift.redhat.com/openshiftmm/listinfo/dev


Problem about logging in openshift origin

2017-09-15 Thread Yu Wei
Hi,

I setup OpenShift origin 3.6 cluster successfully and enabled metrics and 
logging.

Metrics worked well and logging didn't worked.

Pod logging-es-data-master-lf6al5rb-5-deploy in logging frequently crashed with 
below logs,

--> Scaling logging-es-data-master-lf6al5rb-5 to 1
--> Waiting up to 10m0s for pods in rc logging-es-data-master-lf6al5rb-5 to 
become ready
error: update acceptor rejected logging-es-data-master-lf6al5rb-5: pods for rc 
"logging-es-data-master-lf6al5rb-5" took longer than 600 seconds to become ready


I didn't find other information. How could I debug such problem?


Thanks,

Jared, (韦煜)
Software developer
Interested in open source software, big data, Linux
___
dev mailing list
dev@lists.openshift.redhat.com
http://lists.openshift.redhat.com/openshiftmm/listinfo/dev


Re: Problem about logging in openshift origin

2017-09-18 Thread Yu Wei
@Mateus Caruccio

I run the commands you mentioned and did not find any useful information.

It indicated that no pods named logging-es-data-master-lf6al5rb-5.

No event logs found either.


Thanks,

Jared, (韦煜)
Software developer
Interested in open source software, big data, Linux


From: Mateus Caruccio <mateus.caruc...@getupcloud.com>
Sent: Friday, September 15, 2017 6:19:36 PM
To: Yu Wei
Cc: dev@lists.openshift.redhat.com; users
Subject: Re: Problem about logging in openshift origin

You can look into two places for clues.  The pod's log itself (oc -n logging 
logs -f logging-es-data-master-lf6al5rb-5) and project events (oc -n logging 
get events)

Em 15 de set de 2017 07:10, "Yu Wei" 
<yu20...@hotmail.com<mailto:yu20...@hotmail.com>> escreveu:

Hi,

I setup OpenShift origin 3.6 cluster successfully and enabled metrics and 
logging.

Metrics worked well and logging didn't worked.

Pod logging-es-data-master-lf6al5rb-5-deploy in logging frequently crashed with 
below logs,

--> Scaling logging-es-data-master-lf6al5rb-5 to 1
--> Waiting up to 10m0s for pods in rc logging-es-data-master-lf6al5rb-5 to 
become ready
error: update acceptor rejected logging-es-data-master-lf6al5rb-5: pods for rc 
"logging-es-data-master-lf6al5rb-5" took longer than 600 seconds to become ready


I didn't find other information. How could I debug such problem?


Thanks,

Jared, (韦煜)
Software developer
Interested in open source software, big data, Linux

___
users mailing list
us...@lists.openshift.redhat.com<mailto:us...@lists.openshift.redhat.com>
http://lists.openshift.redhat.com/openshiftmm/listinfo/users

___
dev mailing list
dev@lists.openshift.redhat.com
http://lists.openshift.redhat.com/openshiftmm/listinfo/dev


Is that possible to deploy openshift on existing k8s cluster?

2017-08-22 Thread Yu Wei
Hi,

Now we have existing k8s cluster running workloads.

We also want to make use of features provided by Openshift Origin, for example 
DevOps etc.

Is that possible to integrate openshift origin with our existing k8s?


Any advice?


Thanks,

Jared, (韦煜)
Software developer
Interested in open source software, big data, Linux
___
dev mailing list
dev@lists.openshift.redhat.com
http://lists.openshift.redhat.com/openshiftmm/listinfo/dev


Mount failed with dynamic provisioned persistent storage glusterfs

2017-09-27 Thread Yu Wei
Hi,

When trying to provision persistent storage dynamically, below errors was 
caught,

Events:
  FirstSeenLastSeenCountFromSubObjectPathType   
 ReasonMessage
  ---
-----
  5m5m1default-schedulerNormal
ScheduledSuccessfully assigned nginx-3783656783-wv69l to host-10-1-241-54
  5m1m10kubelet, host-10-1-241-54Warning
FailedMountMountVolume.SetUp failed for volume 
"kubernetes.io/glusterfs/ece1a4c4-a364-11e7-9b9b-fa163e3e1b52-pvc-777b1c98-a361-11e7-9b9b-fa163e3e1b52"
 (spec.Name: "pvc-777b1c98-a361-11e7-9b9b-fa163e3e1b52") pod 
"ece1a4c4-a364-11e7-9b9b-fa163e3e1b52" (UID: 
"ece1a4c4-a364-11e7-9b9b-fa163e3e1b52") with: glusterfs: mount failed: exit 
status 1 the following error information was pulled from the glusterfs log to 
help diagnose this issue: glusterfs: could not open log file for pod: 
nginx-3783656783-wv69l
  3m1m2kubelet, host-10-1-241-54Warning
FailedMountUnable to mount volumes for pod 
"nginx-3783656783-wv69l_ai-demo(ece1a4c4-a364-11e7-9b9b-fa163e3e1b52)": timeout 
expired waiting for volumes to attach/mount for pod 
"ai-demo"/"nginx-3783656783-wv69l". list of unattached/unmounted volumes=[html]


If I do not use dynamic provision, it works well.

It seemed that above error was thrown by function readGlusterLog().

https://github.com/openshift/origin/blob/85eb37b34f0657631592356d020cef5a58470f8e/vendor/k8s.io/kubernetes/pkg/volume/glusterfs/glusterfs_util.go


Is there any solution to work around this issue?

Or Did I miss anything?



Thanks,

Jared, (韦煜)
Software developer
Interested in open source software, big data, Linux
___
dev mailing list
dev@lists.openshift.redhat.com
http://lists.openshift.redhat.com/openshiftmm/listinfo/dev


authentication required even for pulling images from private registry

2017-10-18 Thread Yu Wei
Hi,

I setup openshift origin cluster 3.6 and found a problem with private registry.

Image was failed to be pulled by work node with error as below,

rpc error: code = 2 desc = unauthorized: authentication required


However, the registry works well and I also could find the image via 
docker-console.

I installed the cluster via "Advanced installation". It seemed insecure 
registry is not enabled.


How could I check what's wrong in my env?



Thanks,

Jared, (韦煜)
Software developer
Interested in open source software, big data, Linux
___
dev mailing list
dev@lists.openshift.redhat.com
http://lists.openshift.redhat.com/openshiftmm/listinfo/dev


Network issues with openvswitch

2017-10-22 Thread Yu Wei
Hi,

I execute following command on work node of openshift origin cluster 3.6.

[root@host-10-1-130-32 ~]# traceroute docker-registry.default.svc
traceroute to docker-registry.default.svc (172.30.22.28), 30 hops max, 60 byte 
packets
 1  bogon (10.130.2.1)  3005.715 ms !H  3005.682 ms !H  3005.664 ms !H
It seemed content marked in red should be hostname of work node.
How could I debug such issue? Where to start?



Thanks,

Jared, (韦煜)
Software developer
Interested in open source software, big data, Linux
___
dev mailing list
dev@lists.openshift.redhat.com
http://lists.openshift.redhat.com/openshiftmm/listinfo/dev


How could I use oc client to search images in integrated docker-registry?

2017-11-13 Thread Yu Wei
Hi,

How could I list images pushed into integrated docker-registry?

Could oc client be used? Or any other options?


Thanks,

Jared, (韦煜)
Software developer
Interested in open source software, big data, Linux
___
dev mailing list
dev@lists.openshift.redhat.com
http://lists.openshift.redhat.com/openshiftmm/listinfo/dev


Re: How could I use oc client to search images in integrated docker-registry?

2017-11-13 Thread Yu Wei
Hi Graham,


Thanks very much for your help.


Thanks,

Jared, (韦煜)
Software developer
Interested in open source software, big data, Linux


From: Graham Dumpleton <gdump...@redhat.com>
Sent: Monday, November 13, 2017 4:41:00 PM
To: Yu Wei
Cc: us...@lists.openshift.redhat.com; dev@lists.openshift.redhat.com
Subject: Re: How could I use oc client to search images in integrated 
docker-registry?

You can use:

oc get is

to list image streams for images in the registry in your current project.

If you have access to other projects you can use '-n' argument and give the 
name of the project.

Running:

oc get is -n openshift

will list image streams in the default 'openshift' project.

You can use:

oc new-app -L

to list images stream, as well as templates in the current project and the 
default 'openshift' project.

You can use:

oc new-app -S

to search for images amongst image streams in current project, default 
'openshift' project and Docker Hub or other external registry attached with the 
OpenShift cluster.

Graham

On 13 Nov 2017, at 7:32 pm, Yu Wei 
<yu20...@hotmail.com<mailto:yu20...@hotmail.com>> wrote:

Hi,
How could I list images pushed into integrated docker-registry?
Could oc client be used? Or any other options?

Thanks,
Jared, (韦煜)
Software developer
Interested in open source software, big data, Linux
___
users mailing list
us...@lists.openshift.redhat.com<mailto:us...@lists.openshift.redhat.com>
http://lists.openshift.redhat.com/openshiftmm/listinfo/users

___
dev mailing list
dev@lists.openshift.redhat.com
http://lists.openshift.redhat.com/openshiftmm/listinfo/dev


Re: environment variables when running multiple containers in one pod

2017-11-27 Thread Yu Wei
Hi Slava,

The two pvc "ocai-mysql-claim" and "ocai-nb-claim" are consuming two pv which 
are resides on one real storage.  After deploying it on openshift origin 
cluster, "empty" password worked when connecting to mysql.

And endpoint for glusterfs was destroyed soon after deployment, I didn't find 
out the reason.


Then I modified the deployment as below.  And mysql worked as expected.

Now I have two questions:

1, Could two pv reside on one real volume? For example, glusterfs volume

2, I found glusterfs related endpoints disappeared after running deployment 
successfully. How could I debug such issue?


   spec:
  containers:
  - name: ocai
image: aura/web:develop
imagePullPolicy: Always
ports:
- containerPort: 9000
  - name: notebook
image: aura/all-spark-notebook:latest
imagePullPolicy: Always
ports:
- containerPort: 8889
command: ["sh", "-c", "jupyter-notebook --port=8889 
--NotebookApp.token='9e4f96c5239743a8dd5910216c6f02dad89a58932d63db7f' 
--config=/jupyter_notebook_config.py --no-browser"]
volumeMounts:
- mountPath: /data/ai_lab/ocai/
  name: ocai-data
  subPath: nb
resources:
  requests:
memory: 4Gi
  - name: mysql
env:
  - name: MYSQL_ROOT_PASSWORD
value: Ocai@1234
image: aura/mysql:5.7
imagePullPolicy: IfNotPresent
ports:
- containerPort: 3306
volumeMounts:
- mountPath: "/var/lib/mysql"
  name: ocai-data
  subPath: mysql
resources:
  requests:
memory: 2G
  volumes:
  - name: ocai-data
persistentVolumeClaim:
  claimName: ocai-claim


Thanks,

Jared, (韦煜)
Software developer
Interested in open source software, big data, Linux


From: Vyacheslav Semushin <vsemu...@redhat.com>
Sent: Monday, November 27, 2017 9:52:50 PM
To: Yu Wei
Cc: us...@lists.openshift.redhat.com; dev@lists.openshift.redhat.com
Subject: Re: environment variables when running multiple containers in one pod

2017-11-27 10:34 GMT+01:00 Yu Wei 
<yu20...@hotmail.com<mailto:yu20...@hotmail.com>>:

Hi,

When running mysql with other containers within one pod, it seemed that 
environment variable "MYSQL_ROOT_PASSWORD" didn't work as expected.

Password set in "MYSQL_ROOT_PASSWORD" couldn't be used for connecting to mysql.

It doesn't work only when other containers are being created and when they try 
to connect to mysql container? If yes, then it could be a race when mysql 
container isn't ready while others try to connect to it.


--
Slava Semushin | OpenShift
___
dev mailing list
dev@lists.openshift.redhat.com
http://lists.openshift.redhat.com/openshiftmm/listinfo/dev


environment variables when running multiple containers in one pod

2017-11-27 Thread Yu Wei
Hi,

When running mysql with other containers within one pod, it seemed that 
environment variable "MYSQL_ROOT_PASSWORD" didn't work as expected.

Password set in "MYSQL_ROOT_PASSWORD" couldn't be used for connecting to mysql.

But empty password works.

If only running mysql in pod, the variable works fine.


How could I fix the problem?


Detailed deployment file is as below,

spec:
  containers:
  - name: ocai
image: aura/web:develop
imagePullPolicy: Always
ports:
- containerPort: 9000
  - name: notebook
image: aura/all-spark-notebook:latest
imagePullPolicy: Always
ports:
- containerPort: 8889
command: ["sh", "-c", "jupyter-notebook --port=8889 
--NotebookApp.token='9e4f96c5239743a8dd5910216c6f02dad89a58932d63db7f' 
--config=/jupyter_notebook_config.py --no-browser"]
volumeMounts:
- mountPath: /data/ai_lab/ocai/
  name: nb-data
resources:
  requests:
memory: 4Gi
  - name: mysql
env:
  - name: MYSQL_ROOT_PASSWORD
value: Ocai@1234
image: aura/mysql:5.7
imagePullPolicy: IfNotPresent
ports:
- containerPort: 3306
volumeMounts:
- mountPath: "/var/lib/mysql"
  name: mysql-data
resources:
  requests:
memory: 2G
  volumes:
- name: mysql-data
  persistentVolumeClaim:
claimName: ocai-mysql-claim
- name: nb-data
  persistentVolumeClaim:
claimName: ocai-nb-claim





Thanks,

Jared, (韦煜)
Software developer
Interested in open source software, big data, Linux
___
dev mailing list
dev@lists.openshift.redhat.com
http://lists.openshift.redhat.com/openshiftmm/listinfo/dev


question about persistent volume

2017-11-27 Thread Yu Wei
Hi,

Could I create several persistent volume from one real storage volume?


Could several pvc share the same endpoint/svc?


What's the relationship between pvc and endpoint?

In my openshift origin cluster, I found pvc related endpoints were destroyed 
unexpectedly.

How could I find what's happened?


Thanks,

Jared, (韦煜)
Software developer
Interested in open source software, big data, Linux
___
dev mailing list
dev@lists.openshift.redhat.com
http://lists.openshift.redhat.com/openshiftmm/listinfo/dev


Several questions about authorization

2017-12-18 Thread Yu Wei
Hi,

I have several questions about user and authorization management.

1, How could I remove user from project?



[root@host-10-1-236-92 gpu-test]# oc login -u test1 -p test1
Login successful.

You have access to the following projects and can switch between them with 'oc 
project ':

  * aura
test1

Using project "aura".
[root@host-10-1-236-92 gpu-test]# oc project aura
Already on project "aura" on server "https://10.1.241.54:8443;.
[root@host-10-1-236-92 gpu-test]# oc get rolebindings
Error from server (Forbidden): User "test1" cannot list rolebindings in project 
"aura"

How should I remove user "test1" from project "aura"?

And how could I find which users belongs to project "aura"?


2, basic-user

When should "basic-user" be used? It seems that basic-user is cluster wide. 
Is my understanding right?


3, user created automatically

When issues the instructions "oc login -u test2 -p test2", user "test2" is 
to be created automatically.

After user creation, which project does created user belong to?


Thanks,

Jared, (韦煜)
Software developer
Interested in open source software, big data, Linux
___
dev mailing list
dev@lists.openshift.redhat.com
http://lists.openshift.redhat.com/openshiftmm/listinfo/dev


Where could I find api for service catalog?

2017-11-14 Thread Yu Wei
Hi,

I setup openshift origin cluster with service-catalog enabled.

Where could I find api for service catalog?

I checked github and did not find detailed information.


Thanks,

Jared, (韦煜)
Software developer
Interested in open source software, big data, Linux
___
dev mailing list
dev@lists.openshift.redhat.com
http://lists.openshift.redhat.com/openshiftmm/listinfo/dev


question about openshift origin deployer

2017-11-19 Thread Yu Wei
Hi,

How does openshift origin deployer start another container?

I checked docker file about deployer and found stuff as 
"/usr/bin/openshift-deploy"?


How is /usr/bin/openshift-deploy implemented? Does it call docker api?

Is "/usr/bin/openshift-deploy" also open sourced? Where could I find it?


Thanks,

Jared, (韦煜)
Software developer
Interested in open source software, big data, Linux
___
dev mailing list
dev@lists.openshift.redhat.com
http://lists.openshift.redhat.com/openshiftmm/listinfo/dev


答复: question about external load balancer

2018-05-18 Thread Yu Wei
Got that. I will have a look at F5 and netscalers to know more about that.

Thank you very much.

发送自 Windows 10 版邮件<https://go.microsoft.com/fwlink/?LinkId=550986>应用


发件人: Joel Pearson <japear...@agiledigital.com.au>
发送时间: Saturday, May 19, 2018 10:18:34 AM
收件人: Yu Wei
抄送: dev@lists.openshift.redhat.com; us...@lists.openshift.redhat.com
主题: Re: question about external load balancer

OpenShift already has some support for F5 load balancer’s as a router. So maybe 
given the choice between F5 or netscalers, then F5’s might make sense.

But either will work fine, it’s probably more a question of which device you 
have more skills in.

On Wed, 16 May 2018 at 3:17 am, Yu Wei 
<yu20...@hotmail.com<mailto:yu20...@hotmail.com>> wrote:
Hi guys,
I tried to setup openshift origin cluster with multiple masters for HA.
I read the doc in 
https://github.com/redhat-cop/openshift-playbooks/blob/master/playbooks/installation/load_balancing.adoc.

Any other advice for external load balancer?
Which solution should I select for external load balancer?  F5 or netscaler? 
Which is better?
My cluster is about more than 200 physical machines.


Thanks,

Jared, (韦煜)
Software developer
Interested in open source software, big data, Linux

___
users mailing list
us...@lists.openshift.redhat.com<mailto:us...@lists.openshift.redhat.com>
http://lists.openshift.redhat.com/openshiftmm/listinfo/users
___
dev mailing list
dev@lists.openshift.redhat.com
http://lists.openshift.redhat.com/openshiftmm/listinfo/dev


Which branch of ansible playbook should be used when installing openshift origin 3.6?

2017-10-20 Thread Yu Wei
Hi,

I'm a little confused about which branch should be used during "advanced 
installation".

From document in https://github.com/openshift/openshift-ansible,  it seemed 
branch 3.6 should be used.


From doc 

 
https://docs.openshift.org/3.6/install_config/install/host_preparation.html#preparing-for-advanced-installations-origin,
 there is section as below,

Be sure to stay on the master branch of the openshift-ansible repository when 
running an advanced installation.


Which branch should I use during advanced installation?


Please help to clarify this.


Thanks,

Jared, (韦煜)
Software developer
Interested in open source software, big data, Linux
___
dev mailing list
dev@lists.openshift.redhat.com
http://lists.openshift.redhat.com/openshiftmm/listinfo/dev


Re: No route to host when trying to connect to services

2017-10-23 Thread Yu Wei
I didn't resolve this yet.

However, I suspect it's caused by ovs of openstack.

I plan to try again.


Thanks,

Jared, (韦煜)
Software developer
Interested in open source software, big data, Linux


From: Łukasz Strzelec <lukasz.strze...@gmail.com>
Sent: Monday, October 23, 2017 2:32:05 PM
To: Yu Wei
Cc: us...@lists.openshift.redhat.com; dev@lists.openshift.redhat.com
Subject: Re: No route to host when trying to connect to services

Hello mate;)

Did you solve this issue?  I have pretty much the same problem. I noticed that 
when I restart  origin-node.service it helps. But the problem is occuring 
frequently. Sometimes to solve it I have to restart whole docker engine.

I've try to debug this issue,  nothing found. The common issues are also the 
DNS resolving problem. Some times,  pods cannot resolve the svc names.

Best regards

2017-10-20 20:14 GMT+02:00 Yu Wei 
<yu20...@hotmail.com<mailto:yu20...@hotmail.com>>:

Hi guys,

I setup openshift origin cluster 3.6 and deployed 3 zookeeper instances as 
cluster.

I met error “no route to host" when trying to connect to one zookeeper via 
service.

The detailed information is as below,

zookeeper-1   172.30.64.134  
2181/TCP,2888/TCP,3888/TCP   10m
zookeeper-2   172.30.174.48  
2181/TCP,2888/TCP,3888/TCP   10m
zookeeper-3   172.30.223.77  
2181/TCP,2888/TCP,3888/TCP   10m
[root@host-10-1-236-92 ~]# curl -kv zookeeper-1:3888
* Could not resolve host: zookeeper-1; Name or service not known
* Closing connection 0
curl: (6) Could not resolve host: zookeeper-1; Name or service not known
[root@host-10-1-236-92 ~]# curl -kv zookeeper-1.aura.svc:3888
* About to connect() to zookeeper-1.aura.svc port 3888 (#0)
*   Trying 172.30.64.134...
* Connected to zookeeper-1.aura.svc (172.30.64.134) port 3888 (#0)
> GET / HTTP/1.1
> User-Agent: curl/7.29.0
> Host: zookeeper-1.aura.svc:3888
> Accept: */*
>
* Recv failure: Connection reset by peer
* Closing connection 0
curl: (56) Recv failure: Connection reset by peer
[root@host-10-1-236-92 ~]# curl -kv zookeeper-2.aura.svc:3888
* About to connect() to zookeeper-2.aura.svc port 3888 (#0)
*   Trying 172.30.174.48...
* No route to host
* Failed connect to zookeeper-2.aura.svc:3888; No route to host
* Closing connection 0
curl: (7) Failed connect to zookeeper-2.aura.svc:3888; No route to host
[root@host-10-1-236-92 ~]# curl -kv zookeeper-3.aura.svc:3888
* About to connect() to zookeeper-3.aura.svc port 3888 (#0)
*   Trying 172.30.223.77...
* Connected to zookeeper-3.aura.svc (172.30.223.77) port 3888 (#0)
> GET / HTTP/1.1
> User-Agent: curl/7.29.0
> Host: zookeeper-3.aura.svc:3888
> Accept: */*
>
* Recv failure: Connection reset by peer
* Closing connection 0
curl: (56) Recv failure: Connection reset by peer

The pods are running well.
How could I fix such problem?


Thanks,

Jared, (韦煜)
Software developer
Interested in open source software, big data, Linux

___
users mailing list
us...@lists.openshift.redhat.com<mailto:us...@lists.openshift.redhat.com>
http://lists.openshift.redhat.com/openshiftmm/listinfo/users




--
Ł.S.
___
dev mailing list
dev@lists.openshift.redhat.com
http://lists.openshift.redhat.com/openshiftmm/listinfo/dev


Re: Network issues with openvswitch

2017-10-22 Thread Yu Wei
Hi Aleks,


I setup openshift origin cluster with 1lb + 3 masters + 5 nodes.

In some nodes, pods running on them couldn't be reached by other nodes or pods 
running on other nodes. It indicates "no route to host".

[root@host-10-1-130-32 ~]# curl -kv 
docker-registry.default.svc.cluster.local:5000
* About to connect() to docker-registry.default.svc.cluster.local port 5000 (#0)
*   Trying 172.30.22.28...
* No route to host
* Failed connect to docker-registry.default.svc.cluster.local:5000; No route to 
host
* Closing connection 0
curl: (7) Failed connect to docker-registry.default.svc.cluster.local:5000; No 
route to host


And other nodes works fine.

In my previous mail, host name of node is host-10-1-130-32.

Output of "ifconfig tun0" is as below,

[root@host-10-1-130-32 ~]# ifconfig tun0
tun0: flags=4163<UP,BROADCAST,RUNNING,MULTICAST>  mtu 1450
inet 10.130.2.1  netmask 255.255.254.0  broadcast 0.0.0.0
inet6 fe80::cc50:3dff:fe07:9ea2  prefixlen 64  scopeid 0x20
ether ce:50:3d:07:9e:a2  txqueuelen 1000  (Ethernet)
RX packets 97906  bytes 8665783 (8.2 MiB)
RX errors 0  dropped 0  overruns 0  frame 0
TX packets 163379  bytes 27405744 (26.1 MiB)
TX errors 0  dropped 0 overruns 0  carrier 0  collisions 0

I also tried to capture packet via tcpdump, and found some stuff as following,

10.1.130.32.58147 > 10.1.236.92.4789: [no cksum] VXLAN, flags [I] (0x08), vni 0
ARP, Ethernet (len 6), IPv4 (len 4), Request who-has 10.128.1.45 tell 
10.130.2.1, length 28
0x:  04f9 38ae 659b fa16 3e6c dd90 0800 4500  ..8.e...>lE.
0x0010:  004e 543c 4000 4011 63e4 0a01 8220 0a01  .NT<@.@.c...
0x0020:  ec5c e323 12b5 003a  0800    .\.#...:
0x0030:      ce50 3d07 9ea2 0806  .P=.
0x0040:  0001 0800 0604 0001 ce50 3d07 9ea2 0a82  .P=.
0x0050:  0201    0a80 012d...-
   25  00:22:47.214387 ARP, Ethernet (len 6), IPv4 (len 4), Request who-has 
10.1.130.2 tell 10.1.130.45, length 46
0x:     fa16 3e5a a862 0806 0001  >Z.b
0x0010:  0800 0604 0001 fa16 3e5a a862 0a01 822d  >Z.b...-
0x0020:     0a01 8202     
0x0030:       
   26  00:22:47.258344 IP6 (hlim 255, next-header ICMPv6 (58) payload length: 
24) :: > ff02::1:ffa1:1fbb: [icmp6 sum ok] ICMP6, neighbor solicitation, length 
24, who has fe80::824:c2ff:fea1:1fbb
0x:   ffa1 1fbb 0a24 c2a1 1fbb 86dd 6000  33.$..`.
0x0010:   0018 3aff       :...
0x0020:     ff02      
0x0030:  0001 ffa1 1fbb 8700 724a   fe80  rJ..
0x0040:     0824 c2ff fea1 1fbb   ...$..
   27  00:22:47.282619 ARP, Ethernet (len 6), IPv4 (len 4), Request who-has 
10.1.130.2 tell 10.1.130.73, length 46
0x:     fa16 3ec4 a9be 0806 0001  >...
0x0010:  0800 0604 0001 fa16 3ec4 a9be 0a01 8249  >..I
0x0020:     0a01 8202     
0x0030:       

I didn't understand why the IP marked in red above were involved.


Thanks,

Jared, (韦煜)
Software developer
Interested in open source software, big data, Linux


From: Aleksandar Lazic <al...@me2digital.eu>
Sent: Monday, October 23, 2017 2:34:13 AM
To: Yu Wei; us...@lists.openshift.redhat.com; dev@lists.openshift.redhat.com
Subject: Re: Network issues with openvswitch

Hi Yu Wei.

on Sonntag, 22. Oktober 2017 at 19:13 was written:

> Hi,

> I execute following command on work node of openshift origin cluster 3.6.
>
> [root@host-10-1-130-32 ~]# traceroute docker-registry.default.svc
> traceroute to docker-registry.default.svc (172.30.22.28), 30 hops max, 60 
> byte packets
>  1  bogon (10.130.2.1)  3005.715 ms !H  3005.682 ms !H  3005.664 ms !H
>  It seemed content marked in red  should be hostname of work node.
>  How could I debug such issue? Where to start?

What's the hostname of the node?
I'm not sure what you try to debug or what's the problem you try to
solve?

> Thanks,

> Jared, (韦煜)
>  Software developer
>  Interested in open source software, big data, Linux

--
Best Regards
Aleks
___
dev mailing list
dev@lists.openshift.redhat.com
http://lists.openshift.redhat.com/openshiftmm/listinfo/dev


How to deploy openshift origin cluster on openstack?

2018-04-10 Thread Yu Wei
Hi,
How to deploy openshift origin cluster on openstack?
Could I use magnum, heat or other components?

Is there any document about this?


Thanks,

Jared, (韦煜)
Software developer
Interested in open source software, big data, Linux
___
dev mailing list
dev@lists.openshift.redhat.com
http://lists.openshift.redhat.com/openshiftmm/listinfo/dev


Re: Install OpenShift Origin 3.9 failed on single node

2018-04-11 Thread Yu Wei
In my previous installation, "openshift_release" and "openshift_pkg_version" 
were set to 3.9 and installation failed.
I changed the value to v3.9, it worked now.
When I installed OpenShift Origin 3.6, the value was set to 3.6 without prefix 
'v'.

Is there any changes?

Thanks again for help.

From: Charles Moulliard <cmoul...@redhat.com>
Sent: Tuesday, April 10, 2018 8:24 PM
To: Clayton Coleman
Cc: Yu Wei; us...@lists.openshift.redhat.com; dev@lists.openshift.redhat.com
Subject: Re: Install OpenShift Origin 3.9 failed on single node

I think that something is not well defined under your inventory as this error 
arrives due to the following reason

def lib_utils_oo_image_tag_to_rpm_version(version, include_dash=False):
""" Convert an image tag string to an RPM version if necessary
Empty strings and strings that are already in rpm version format
are ignored. Also remove non semantic version components.

Ex. v3.2.0.10 -> -3.2.0.10
v1.2.0-rc1 -> -1.2.0
"""
if not isinstance(version, string_types):
raise errors.AnsibleFilterError("|failed expects a string or unicode")



CHARLES MOULLIARD

SOFTWARE ENGINEER MANAGER SPRING(BOOT)

Red Hat<https://www.redhat.com/>

cmoulli...@redhat.com<mailto:cmoulli...@redhat.com>M: 
+32-473-604014<tel:+32-473-604014>

[https://www.redhat.com/files/brand/email/sig-redhat.png]<https://red.ht/sig>
@cmoulliard<https://twitter.com/cmoulliard>

On Tue, Apr 10, 2018 at 2:17 PM, Clayton Coleman 
<ccole...@redhat.com<mailto:ccole...@redhat.com>> wrote:
You can try rerunning the install with -vv to get additional debug information.

What OS and version on Ansible are you using?

On Apr 10, 2018, at 3:24 AM, Yu Wei 
<yu20...@hotmail.com<mailto:yu20...@hotmail.com>> wrote:

Hi,
I tried to install openshift origin 3.9 on a single machine and encountered 
problems as below,

TASK [openshift_node : Install Node package, sdn-ovs, conntrack packages] 
*
fatal: [host-10-1-241-74]: FAILED! => {"msg": "|failed expects a string or 
unicode"}
to retry, use: --limit 
@/root/jared/openshift-ansible/playbooks/deploy_cluster.retry

PLAY RECAP 

host-10-1-241-74   : ok=326  changed=41   unreachable=0failed=1
localhost  : ok=13   changed=0unreachable=0failed=0


INSTALLER STATUS 
**
Initialization : Complete (0:00:43)
Health Check   : Complete (0:00:05)
etcd Install   : Complete (0:00:58)
Master Install : Complete (0:05:03)
Master Additional Install  : Complete (0:00:48)
Node Install   : In Progress (0:00:38)
This phase can be restarted by running: playbooks/openshift-node/config.yml



Failure summary:


  1. Hosts:host-10-1-241-74
 Play: Configure containerized nodes
 Task: Install Node package, sdn-ovs, conntrack packages
 Message:  |failed expects a string or unicode

I didn't find useful information in docker / journal logs.
How could I fix this problem further?


Thanks,

Jared, (韦煜)
Software developer
Interested in open source software, big data, Linux

___
dev mailing list
dev@lists.openshift.redhat.com<mailto:dev@lists.openshift.redhat.com>
http://lists.openshift.redhat.com/openshiftmm/listinfo/dev

___
users mailing list
us...@lists.openshift.redhat.com<mailto:us...@lists.openshift.redhat.com>
http://lists.openshift.redhat.com/openshiftmm/listinfo/users


___
dev mailing list
dev@lists.openshift.redhat.com
http://lists.openshift.redhat.com/openshiftmm/listinfo/dev


Install OpenShift Origin 3.9 failed on single node

2018-04-10 Thread Yu Wei
Hi,
I tried to install openshift origin 3.9 on a single machine and encountered 
problems as below,

TASK [openshift_node : Install Node package, sdn-ovs, conntrack packages] 
*
fatal: [host-10-1-241-74]: FAILED! => {"msg": "|failed expects a string or 
unicode"}
to retry, use: --limit 
@/root/jared/openshift-ansible/playbooks/deploy_cluster.retry

PLAY RECAP 

host-10-1-241-74   : ok=326  changed=41   unreachable=0failed=1
localhost  : ok=13   changed=0unreachable=0failed=0


INSTALLER STATUS 
**
Initialization : Complete (0:00:43)
Health Check   : Complete (0:00:05)
etcd Install   : Complete (0:00:58)
Master Install : Complete (0:05:03)
Master Additional Install  : Complete (0:00:48)
Node Install   : In Progress (0:00:38)
This phase can be restarted by running: playbooks/openshift-node/config.yml



Failure summary:


  1. Hosts:host-10-1-241-74
 Play: Configure containerized nodes
 Task: Install Node package, sdn-ovs, conntrack packages
 Message:  |failed expects a string or unicode

I didn't find useful information in docker / journal logs.
How could I fix this problem further?


Thanks,

Jared, (韦煜)
Software developer
Interested in open source software, big data, Linux
___
dev mailing list
dev@lists.openshift.redhat.com
http://lists.openshift.redhat.com/openshiftmm/listinfo/dev


when delete deployment via rest api, corresponding rs and pod not deleted

2018-12-25 Thread Yu Wei
Hi,

By calling rest api to delete deployment, deployment was deleted 
successfully, however, corresponding replica set and pods were not deleted.
How could I delete deployment via rest api?

-- 
Thanks
Jared
Interested in big data, cloud computing


___
dev mailing list
dev@lists.openshift.redhat.com
http://lists.openshift.redhat.com/openshiftmm/listinfo/dev


Could supplemental Groups be used for ceph rbd?

2019-03-30 Thread Yu Wei
Hi,

I setup ceph rbd with openshift origin 3.6.

I read document about ceph volume and volume security.
https://docs.okd.io/3.6/install_config/persistent_storage/pod_security_context.html#supplemental-groups
https://docs.okd.io/3.6/install_config/persistent_storage/persistent_storage_ceph_rbd.html

It seemed that for ceph rbd,  fsGroups rather than supplemental groups 
should be used.

In my test, I didn't specify fsGroup for pods. But it worked.  It seemed 
that supplemental groups was used as file groups for mounted volumes.

I also run the same test in anther testing environment. Pods failed to 
start and reported permission issues for the mounted volume. However, pv 
and pvc were created successfully.

I have two questions as below,
1, Could only fsGroup be used for ceph rbd? Could supplemental groups 
also be used?
2, Behavior is different in my two test environments. One succeed and 
one failed.
     Except supplemental groups and fsGroups, is there any other 
settings that could affect volume access?

-- 
Thanks,
Jared
Interested in big data, cloud computing


___
dev mailing list
dev@lists.openshift.redhat.com
http://lists.openshift.redhat.com/openshiftmm/listinfo/dev


How could I config calico for openshift-sdn?

2019-09-03 Thread Yu Wei
Hi guys,
I want to setup old cluster with calico plugin.
How could I configure it?
Is there any doc about this?

Thanks,
Jared


___
dev mailing list
dev@lists.openshift.redhat.com
http://lists.openshift.redhat.com/openshiftmm/listinfo/dev


Re: Pods stuck on 'ContainerCreating' when redhat/openshift-ovs-multitenant enabled

2019-10-15 Thread Yu Wei
I found the root cause for this issue.
In my machine, I firstly deployed cop with calico. It works well.
Then run uninstall playbook and reinstall with sdn openshift-ovs-multitenant.
And it didn’t work anymore.
I found something as below,

[root@buzz1 openshift-ansible]# systemctl status  atomic-openshift-node.service
● atomic-openshift-node.service - OpenShift Node
   Loaded: loaded (/etc/systemd/system/atomic-openshift-node.service; enabled; 
vendor preset: disabled)
   Active: active (running) since Mon 2019-10-14 00:43:08 PDT; 22h ago
 Docs: https://github.com/openshift/origin
 Main PID: 87388 (hyperkube)
   CGroup: /system.slice/atomic-openshift-node.service
   ├─87388 /usr/bin/hyperkube kubelet --v=6 --address=0.0.0.0 
--allow-privileged=true --anonymous-auth=true --authentication-toke...
   └─88872 /opt/cni/bin/calico

Oct 14 23:15:48 buzz1.fyre.ibm.com<http://buzz1.fyre.ibm.com> 
atomic-openshift-node[87388]: I1014 23:15:48.289674   87388 common.go:71] Using 
namespace "kube-syaml
Oct 14 23:15:48 buzz1.fyre.ibm.com<http://buzz1.fyre.ibm.com> 
atomic-openshift-node[87388]: I1014 23:15:48.289809   87388 file.go:199] 
Reading config file "/et...yaml"
Oct 14 23:15:48 buzz1.fyre.ibm.com<http://buzz1.fyre.ibm.com> 
atomic-openshift-node[87388]: I1014 23:15:48.292556   87388 common.go:62] 
Generated UID "598eab3cyaml
Oct 14 23:15:48 buzz1.fyre.ibm.com<http://buzz1.fyre.ibm.com> 
atomic-openshift-node[87388]: I1014 23:15:48.293602   87388 common.go:66] 
Generated Name "master-yaml
Oct 14 23:15:48 buzz1.fyre.ibm.com<http://buzz1.fyre.ibm.com> 
atomic-openshift-node[87388]: I1014 23:15:48.294512   87388 common.go:71] Using 
namespace "kube-syaml
Oct 14 23:15:48 buzz1.fyre.ibm.com<http://buzz1.fyre.ibm.com> 
atomic-openshift-node[87388]: I1014 23:15:48.295667   87388 file.go:199] 
Reading config file "/et...yaml"
Oct 14 23:15:48 buzz1.fyre.ibm.com<http://buzz1.fyre.ibm.com> 
atomic-openshift-node[87388]: I1014 23:15:48.296350   87388 common.go:62] 
Generated UID "d71dc810yaml
Oct 14 23:15:48 buzz1.fyre.ibm.com<http://buzz1.fyre.ibm.com> 
atomic-openshift-node[87388]: I1014 23:15:48.296367   87388 common.go:66] 
Generated Name "master-yaml
Oct 14 23:15:48 buzz1.fyre.ibm.com<http://buzz1.fyre.ibm.com> 
atomic-openshift-node[87388]: I1014 23:15:48.296379   87388 common.go:71] Using 
namespace "kube-syaml
Oct 14 23:15:48 buzz1.fyre.ibm.com<http://buzz1.fyre.ibm.com> 
atomic-openshift-node[87388]: I1014 23:15:48.300194   87388 config.go:303] 
Setting pods for source file
Oct 14 23:15:48 buzz1.fyre.ibm.com<http://buzz1.fyre.ibm.com> 
atomic-openshift-node[87388]: I1014 23:15:48.361625   87388 kubelet.go:1884] 
SyncLoop (SYNC): 3 p...d33c)
Oct 14 23:15:48 buzz1.fyre.ibm.com<http://buzz1.fyre.ibm.com> 
atomic-openshift-node[87388]: I1014 23:15:48.361693   87388 config.go:100] 
Looking for [api file]...e:{}]
Oct 14 23:15:48 buzz1.fyre.ibm.com<http://buzz1.fyre.ibm.com> 
atomic-openshift-node[87388]: I1014 23:15:48.361716   87388 kubelet.go:1907] 
SyncLoop (housekeeping)
Hint: Some lines were ellipsized, use -l to show in full.
[root@buzz1 openshift-ansible]# ps -ef | grep calico
root  88872  87388  0 23:15 ?00:00:00 /opt/cni/bin/calico
root  88975  74601  0 23:15 pts/000:00:00 grep --color=auto calico
[root@buzz1 openshift-ansible]#

It seemed that calico is extra here. Then using the same inventory file, OCP 
3.11 could be deployed on a clean VM successfully.
I guessed that uninstall playbook did not clear calico thoroughly.


On Oct 12, 2019, at 11:52 PM, Yu Wei 
mailto:yu20...@hotmail.com>> wrote:

Hi,
I tried to install OCP 3.11 with following variables set.
openshift_use_openshift_sdn=true
os_sdn_network_plugin_name='redhat/openshift-ovs-multitenant’

Some pods stuck on ‘ContainerCreating’.
[root@buzz1 openshift-ansible]# oc get pods --all-namespaces
NAMESPACE   NAMEREADY 
STATUS  RESTARTS   AGE
default docker-registry-1-deploy0/1   
ContainerCreating   0  5h
default registry-console-1-deploy   0/1   
ContainerCreating   0  5h
kube-system 
master-api-buzz1.center1.com<http://master-api-buzz1.center1.com/>
1/1   Running 0  5h
kube-system 
master-controllers-buzz1.center1.com<http://master-controllers-buzz1.center1.com/>
1/1   Running 0  5h
kube-system 
master-etcd-buzz1.center1.com<http://master-etcd-buzz1.center1.com/>   
1/1   Running 0  5h
openshift-node  sync-x8j7d  1/1   
Running 0  5h
openshift-sdn   ovs-ff7r7   

Re: Pods stuck on 'ContainerCreating' when redhat/openshift-ovs-multitenant enabled

2019-10-16 Thread Yu Wei
Hi Dan,
I checked the logs of all pods in namespace openshift-sdn and I didn’t find any 
errors in them.
I reinstalled with ‘redhat/openshift-ovs-multitenant’ on a clean machine, 
everything works well.

So I suspect uninstall playbook didn’t clean calico plugin properly.

Thanks,
Jared


On Oct 16, 2019, at 1:09 AM, Dan Williams 
mailto:d...@redhat.com>> wrote:

On Tue, 2019-10-15 at 06:18 +0000, Yu Wei wrote:
I found the root cause for this issue.
In my machine, I firstly deployed cop with calico. It works well.
Then run uninstall playbook and reinstall with sdn openshift-ovs-
multitenant.
And it didn’t work anymore.
I found something as below,

[root@buzz1 openshift-ansible]# systemctl status  atomic-openshift-
node.service
● atomic-openshift-node.service - OpenShift Node
  Loaded: loaded (/etc/systemd/system/atomic-openshift-node.service;
enabled; vendor preset: disabled)
  Active: active (running) since Mon 2019-10-14 00:43:08 PDT; 22h
ago
Docs: https://github.com/openshift/origin
Main PID: 87388 (hyperkube)
  CGroup: /system.slice/atomic-openshift-node.service
  ├─87388 /usr/bin/hyperkube kubelet --v=6 --address=0.0.0.0
--allow-privileged=true --anonymous-auth=true --authentication-
toke...
  └─88872 /opt/cni/bin/calico

Oct 14 23:15:48 
buzz1.fyre.ibm.com<http://buzz1.fyre.ibm.com/><http://buzz1.fyre.ibm.com<http://buzz1.fyre.ibm.com/>>
 atomic-
openshift-node[87388]: I1014 23:15:48.289674   87388 common.go:71]
Using namespace "kube-syaml
Oct 14 23:15:48 
buzz1.fyre.ibm.com<http://buzz1.fyre.ibm.com/><http://buzz1.fyre.ibm.com<http://buzz1.fyre.ibm.com/>>
 atomic-
openshift-node[87388]: I1014 23:15:48.289809   87388 file.go:199]
Reading config file "/et...yaml"
Oct 14 23:15:48 
buzz1.fyre.ibm.com<http://buzz1.fyre.ibm.com/><http://buzz1.fyre.ibm.com<http://buzz1.fyre.ibm.com/>>
 atomic-
openshift-node[87388]: I1014 23:15:48.292556   87388 common.go:62]
Generated UID "598eab3cyaml
Oct 14 23:15:48 
buzz1.fyre.ibm.com<http://buzz1.fyre.ibm.com/><http://buzz1.fyre.ibm.com<http://buzz1.fyre.ibm.com/>>
 atomic-
openshift-node[87388]: I1014 23:15:48.293602   87388 common.go:66]
Generated Name "master-yaml
Oct 14 23:15:48 
buzz1.fyre.ibm.com<http://buzz1.fyre.ibm.com/><http://buzz1.fyre.ibm.com<http://buzz1.fyre.ibm.com/>>
 atomic-
openshift-node[87388]: I1014 23:15:48.294512   87388 common.go:71]
Using namespace "kube-syaml
Oct 14 23:15:48 
buzz1.fyre.ibm.com<http://buzz1.fyre.ibm.com/><http://buzz1.fyre.ibm.com<http://buzz1.fyre.ibm.com/>>
 atomic-
openshift-node[87388]: I1014 23:15:48.295667   87388 file.go:199]
Reading config file "/et...yaml"
Oct 14 23:15:48 
buzz1.fyre.ibm.com<http://buzz1.fyre.ibm.com/><http://buzz1.fyre.ibm.com<http://buzz1.fyre.ibm.com/>>
 atomic-
openshift-node[87388]: I1014 23:15:48.296350   87388 common.go:62]
Generated UID "d71dc810yaml
Oct 14 23:15:48 
buzz1.fyre.ibm.com<http://buzz1.fyre.ibm.com/><http://buzz1.fyre.ibm.com<http://buzz1.fyre.ibm.com/>>
 atomic-
openshift-node[87388]: I1014 23:15:48.296367   87388 common.go:66]
Generated Name "master-yaml
Oct 14 23:15:48 
buzz1.fyre.ibm.com<http://buzz1.fyre.ibm.com/><http://buzz1.fyre.ibm.com<http://buzz1.fyre.ibm.com/>>
 atomic-
openshift-node[87388]: I1014 23:15:48.296379   87388 common.go:71]
Using namespace "kube-syaml
Oct 14 23:15:48 
buzz1.fyre.ibm.com<http://buzz1.fyre.ibm.com/><http://buzz1.fyre.ibm.com<http://buzz1.fyre.ibm.com/>>
 atomic-
openshift-node[87388]: I1014 23:15:48.300194   87388 config.go:303]
Setting pods for source file
Oct 14 23:15:48 
buzz1.fyre.ibm.com<http://buzz1.fyre.ibm.com/><http://buzz1.fyre.ibm.com<http://buzz1.fyre.ibm.com/>>
 atomic-
openshift-node[87388]: I1014 23:15:48.361625   87388 kubelet.go:1884]
SyncLoop (SYNC): 3 p...d33c)
Oct 14 23:15:48 
buzz1.fyre.ibm.com<http://buzz1.fyre.ibm.com/><http://buzz1.fyre.ibm.com<http://buzz1.fyre.ibm.com/>>
 atomic-
openshift-node[87388]: I1014 23:15:48.361693   87388 config.go:100]
Looking for [api file]...e:{}]
Oct 14 23:15:48 
buzz1.fyre.ibm.com<http://buzz1.fyre.ibm.com/><http://buzz1.fyre.ibm.com<http://buzz1.fyre.ibm.com/>>
 atomic-
openshift-node[87388]: I1014 23:15:48.361716   87388 kubelet.go:1907]
SyncLoop (housekeeping)
Hint: Some lines were ellipsized, use -l to show in full.
[root@buzz1 openshift-ansible]# ps -ef | grep calico
root  88872  87388  0 23:15 ?00:00:00 /opt/cni/bin/calico
root  88975  74601  0 23:15 pts/0    00:00:00 grep --color=auto
calico
[root@buzz1 openshift-ansible]#

It seemed that calico is extra here. Then using the same inventory
file, OCP 3.11 could be deployed on a clean VM successfully.
I guessed that uninstall playbook did not clear calico thoroughly.


On 

master api failed to connect to etcd when installing open shift-enterprise 3.11

2019-09-25 Thread Yu Wei
Hi,
I tried to install openshift enterprise 3.11 on one single node with two NICs.
Etcd and master-controller started and master-api failed to start due to 
connection to etcd refused.

Log from master-api container as below,
I0924 22:58:20.297828   1 plugins.go:84] Registered admission plugin 
"NamespaceExists"
I0924 22:58:20.297835   1 plugins.go:84] Registered admission plugin 
"NodeRestriction"
I0924 22:58:20.297844   1 plugins.go:84] Registered admission plugin 
"PersistentVolumeLabel"
I0924 22:58:20.297857   1 plugins.go:84] Registered admission plugin 
"PodNodeSelector"
I0924 22:58:20.297866   1 plugins.go:84] Registered admission plugin 
"PodPreset"
I0924 22:58:20.297875   1 plugins.go:84] Registered admission plugin 
"PodTolerationRestriction"
I0924 22:58:20.297885   1 plugins.go:84] Registered admission plugin 
"ResourceQuota"
I0924 22:58:20.297894   1 plugins.go:84] Registered admission plugin 
"PodSecurityPolicy"
I0924 22:58:20.297901   1 plugins.go:84] Registered admission plugin 
"Priority"
I0924 22:58:20.297908   1 plugins.go:84] Registered admission plugin 
"SecurityContextDeny"
I0924 22:58:20.297927   1 plugins.go:84] Registered admission plugin 
"ServiceAccount"
I0924 22:58:20.297934   1 plugins.go:84] Registered admission plugin 
"DefaultStorageClass"
I0924 22:58:20.297941   1 plugins.go:84] Registered admission plugin 
"PersistentVolumeClaimResize"
I0924 22:58:20.297948   1 plugins.go:84] Registered admission plugin 
"StorageObjectInUseProtection"
F0924 22:58:50.301065   1 start_api.go:68] dial tcp 172.16.211.60:2379: 
connect: connection refused

Log from etcd as below,
019-09-24 20:34:25.292743 I | embed: ClientTLS: cert = /etc/etcd/server.crt, 
key = /etc/etcd/server.key, ca = , trusted-ca = /etc/etcd/ca.crt, 
client-cert-auth = true
2019-09-24 20:34:25.293054 I | etcdserver/membership: added member 
9b85c8d12305d256 [https://9.30.119.88:2380] to cluster 1d7f9f2198813e8c
2019-09-24 20:34:26.272070 I | raft: 9b85c8d12305d256 is starting a new 
election at term 1
2019-09-24 20:34:26.272167 I | raft: 9b85c8d12305d256 became candidate at term 2
2019-09-24 20:34:26.272206 I | raft: 9b85c8d12305d256 received MsgVoteResp from 
9b85c8d12305d256 at term 2
2019-09-24 20:34:26.272226 I | raft: 9b85c8d12305d256 became leader at term 2
2019-09-24 20:34:26.272234 I | raft: raft.node: 9b85c8d12305d256 elected leader 
9b85c8d12305d256 at term 2
2019-09-24 20:34:26.272617 I | etcdserver: published 
{Name:buzz1.fyre.ibm.com 
ClientURLs:[https://9.30.119.88:2379]} to cluster 1d7f9f2198813e8c
2019-09-24 20:34:26.272734 I | etcdserver: setting up the initial cluster 
version to 3.2
2019-09-24 20:34:26.272774 I | embed: ready to serve client requests
2019-09-24 20:34:26.273020 I | embed: serving client requests on 
9.30.119.88:2379
2019-09-24 20:34:26.274011 N | etcdserver/membership: set the initial cluster 
version to 3.2
2019-09-24 20:34:26.274069 I | etcdserver/api: enabled capabilities for version 
3.2

Output from ‘ifconfig -a’ as below,
[root@buzz1 openshift-ansible]# ifconfig -a
docker0: flags=4099  mtu 1500
inet 172.17.0.1  netmask 255.255.0.0  broadcast 0.0.0.0
ether 02:42:d0:74:98:c6  txqueuelen 0  (Ethernet)
RX packets 0  bytes 0 (0.0 B)
RX errors 0  dropped 0  overruns 0  frame 0
TX packets 0  bytes 0 (0.0 B)
TX errors 0  dropped 0 overruns 0  carrier 0  collisions 0

eth0: flags=4163  mtu 1500
inet 172.16.211.60  netmask 255.255.0.0  broadcast 172.16.255.255
ether 00:16:3e:01:d3:3c  txqueuelen 1000  (Ethernet)
RX packets 19994481  bytes 2780261854 (2.5 GiB)
RX errors 0  dropped 1  overruns 0  frame 0
TX packets 3220  bytes 285148 (278.4 KiB)
TX errors 0  dropped 0 overruns 0  carrier 0  collisions 0

eth1: flags=4163  mtu 1500
inet 9.30.119.88  netmask 255.255.252.0  broadcast 9.30.119.255
ether 00:00:09:1e:77:58  txqueuelen 1000  (Ethernet)
RX packets 308744  bytes 319369670 (304.5 MiB)
RX errors 0  dropped 6  overruns 0  frame 0
TX packets 124636  bytes 15097082 (14.3 MiB)
TX errors 0  dropped 0 overruns 0  carrier 0  collisions 0

lo: flags=73  mtu 65536
inet 127.0.0.1  netmask 255.0.0.0
loop  txqueuelen 1000  (Local Loopback)
RX packets 423233  bytes 216825068 (206.7 MiB)
RX errors 0  dropped 0  overruns 0  frame 0
TX packets 423233  bytes 216825068 (206.7 MiB)
TX errors 0  dropped 0 overruns 0  carrier 0  collisions 0


Is the issue caused by two NICs?

Thanks,
Jared
___
dev mailing list
dev@lists.openshift.redhat.com
http://lists.openshift.redhat.com/openshiftmm/listinfo/dev


Re: master api failed to connect to etcd when installing open shift-enterprise 3.11

2019-09-25 Thread Yu Wei
Toni, thanks for your help.
Yes, 9.30.x.x is the default route of VM as it’s external IP. 
And 172.16.x.x is internal IP.

How could I resolve the problem? 
Is it possible to make master-api and etcd to use same IP address by specifying 
something in inventory file?


> On Sep 25, 2019, at 5:16 PM, Toni Schmidbauer 
>  wrote:
> 
> 
> Yu Wei  writes:
>> F0924 22:58:50.301065   1 start_api.go:68] dial tcp 172.16.211.60:2379: 
>> connect: connection refused
> 
> the api tries to contact etcd on 172.16.211.60:2379 but
> 
>> 2019-09-24 20:34:26.273020 I | embed: serving client requests on 
>> 9.30.119.88:2379
> 
> but etcd listens for client request on 9.30.119.88.
> 
> a wild guess is that your hostname resolves to 172.16.x.x but your
> default route points to 9.30.x.x.
> 
> for a longer explanation why i think this is the case read on :-)
> 
> the listing address for etcd is configure in /etc/etcd/etcd.conf
> (ETCD_LISTEN_CLIENT_URLS).
> 
> etcd.conf gets created by the openshift-ansible role etcd
> (/usr/share/ansible/openshift-ansible/roles/etcd/templates/etcd.conf.j2),
> it uses etcd_listen_client_urls.
> 
> etcd_listen_client_urls uses etcd_ip which gets set via
> openshift.common.ip, that's a custom fact to be found in
> openshift_facts.py
> (.../roles/openshift_facts/library/openshift_facts.py).
> 
> in get_defaults() ip_addr gets set to the standard ansible fact
> ansible_default_ipv4.address, this is then used common.ip.
> 
> so it seems that ansible thinks that 9.3.30.119.88 is your default
> interface, i'm not 100% sure, but this might be because your default
> route points to this nic (ip route sh).
> 
> the config option for the api to contact etcd is in
> /etc/origin/master/master-config.yaml, etcdClientInfo.urls.
> openshift-ansible uses openshift_master_etcd_hosts for this, afaik this
> is set via the custom fact openshift.common.hostname (once again in
> openshift_facts.py) for figuring out the public hostname (it reuses
> ansible_nodename and ansible_fqdn).
> 
> hopefully this is not too confusing...
> 
> hth
> toni
> 
> ___
> users mailing list
> us...@lists.openshift.redhat.com
> http://lists.openshift.redhat.com/openshiftmm/listinfo/users


___
dev mailing list
dev@lists.openshift.redhat.com
http://lists.openshift.redhat.com/openshiftmm/listinfo/dev