Could you try the following: on the vms, lower the mtu of the vnics connected to the ovn network? And try again?
On Tue, May 8, 2018 at 11:40 AM, Samuli Heinonen <[email protected]> wrote: > Hi Marcin, > > Here is ip addr output from virtual machines: > > [root@testi2 ~]# ip addr > 1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN qlen 1 > link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 > inet 127.0.0.1/8 scope host lo > valid_lft forever preferred_lft forever > inet6 ::1/128 scope host > valid_lft forever preferred_lft forever > 2: eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast > state UP qlen 1000 > link/ether 00:1a:4a:16:01:05 brd ff:ff:ff:ff:ff:ff > inet 10.0.1.25/24 brd 10.0.1.255 scope global dynamic eth0 > valid_lft 86331sec preferred_lft 86331sec > inet6 fe80::21a:4aff:fe16:105/64 scope link > valid_lft forever preferred_lft forever > 3: eth2: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast > state UP qlen 1000 > link/ether 00:1a:4a:16:01:03 brd ff:ff:ff:ff:ff:ff > inet 10.0.200.10/24 brd 10.0.200.255 scope global dynamic eth2 > valid_lft 86334sec preferred_lft 86334sec > inet6 fe80::21a:4aff:fe16:103/64 scope link > valid_lft forever preferred_lft forever > > eth0 connected to network ovirtmgmt > eth2 connected to OVN network vm-public > > [root@testi6 ~]# ip addr > 1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN qlen 1 > link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 > inet 127.0.0.1/8 scope host lo > valid_lft forever preferred_lft forever > inet6 ::1/128 scope host > valid_lft forever preferred_lft forever > 2: eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast > state UP qlen 1000 > link/ether 00:1a:4a:16:01:0b brd ff:ff:ff:ff:ff:ff > inet 10.0.1.27/24 brd 10.0.1.255 scope global dynamic eth0 > valid_lft 86187sec preferred_lft 86187sec > inet6 fe80::21a:4aff:fe16:10b/64 scope link > valid_lft forever preferred_lft forever > 3: eth1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast > state UP qlen 1000 > link/ether 00:1a:4a:16:01:0c brd ff:ff:ff:ff:ff:ff > inet 10.0.200.11/24 brd 10.0.200.255 scope global dynamic eth1 > valid_lft 86301sec preferred_lft 86301sec > inet6 fe80::21a:4aff:fe16:10c/64 scope link > valid_lft forever preferred_lft forever > > eth0 connected to network ovirtmgmt > eth1 connected to OVN network vm-public > > Best regards, > Samuli > > > > Marcin Mirecki kirjoitti 08.05.2018 10:14: > >> Hi Samuli, >> >> Your configuration looks correct. >> Can you also send me the result of 'ip addr' on your vm's? >> >> Thanks, >> Marcin >> >> On Mon, May 7, 2018 at 7:44 PM, Samuli Heinonen >> <[email protected]> wrote: >> >> Hi Marcin, >>> >>> Thank you for your response. >>> >>> I used engine-setup to do the configuration. Only exception is that >>> I had to run "vdsm-tool ovn-config engine-ip local-ip" (ie. >>> vdsm-tool ovn-config 10.0.1.101 10.0.1.21) on hypervisors. >>> >>> Here is the output of requested commands: >>> >>> [root@oe ~]# ovn-sbctl show >>> Chassis "049183d5-61b6-4b9c-bae3-c7b10d30f8cb" >>> hostname: "o2.hirundinidae.local" >>> Encap geneve >>> ip: "10.0.1.18" >>> options: {csum="true"} >>> Port_Binding "87c5e44a-7c8b-41b2-89a6-fa52f27643ed" >>> Chassis "972f1b7b-10de-4e4f-a5f9-f080890f087d" >>> hostname: "o3.hirundinidae.local" >>> Encap geneve >>> ip: "10.0.1.21" >>> options: {csum="true"} >>> Port_Binding "ccea5185-3efa-4d9c-9475-9e46009fea4f" >>> Port_Binding "e868219c-f16c-45c6-b7b1-72d044fee602" >>> >>> [root@oe ~]# ovn-nbctl show >>> switch 7d264a6c-ea48-4a6d-9663-5244102dc9bb (vm-private) >>> port 4ec3ecf6-d04a-406c-8354-c5e195ffde05 >>> addresses: ["00:1a:4a:16:01:06 dynamic"] >>> switch 40aedb7d-b1c3-400e-9ddb-16bee3bb312a (vm-public) >>> port 87c5e44a-7c8b-41b2-89a6-fa52f27643ed >>> addresses: ["00:1a:4a:16:01:03"] >>> port ccea5185-3efa-4d9c-9475-9e46009fea4f >>> addresses: ["00:1a:4a:16:01:0c"] >>> port e868219c-f16c-45c6-b7b1-72d044fee602 >>> addresses: ["00:1a:4a:16:01:0a"] >>> >>> [root@o2 ~]# ip addr >>> 1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN >>> qlen 1 >>> link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 >>> inet 127.0.0.1/8 [1] scope host lo >>> valid_lft forever preferred_lft forever >>> inet6 ::1/128 scope host >>> valid_lft forever preferred_lft forever >>> 2: enp0s31f6: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc >>> pfifo_fast master ovirtmgmt state UP qlen 1000 >>> link/ether 78:f2:9e:90:bc:64 brd ff:ff:ff:ff:ff:ff >>> 3: enp0s20f0u5c2: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc >>> pfifo_fast master public state UNKNOWN qlen 1000 >>> link/ether 50:3e:aa:4c:9b:01 brd ff:ff:ff:ff:ff:ff >>> 4: ovs-system: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN >>> qlen 1000 >>> link/ether 82:49:e1:15:af:56 brd ff:ff:ff:ff:ff:ff >>> 5: br-int: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN qlen >>> 1000 >>> link/ether a2:bb:78:7e:35:4b brd ff:ff:ff:ff:ff:ff >>> 21: public: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue >>> state UP qlen 1000 >>> link/ether 50:3e:aa:4c:9b:01 brd ff:ff:ff:ff:ff:ff >>> inet6 fe80::523e:aaff:fe4c:9b01/64 scope link >>> valid_lft forever preferred_lft forever >>> 22: ovirtmgmt: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc >>> noqueue state UP qlen 1000 >>> link/ether 78:f2:9e:90:bc:64 brd ff:ff:ff:ff:ff:ff >>> inet 10.0.1.18/24 [2] brd 10.0.1.255 scope global ovirtmgmt >>> >>> valid_lft forever preferred_lft forever >>> inet6 fe80::7af2:9eff:fe90:bc64/64 scope link >>> valid_lft forever preferred_lft forever >>> 23: genev_sys_6081: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 65000 >>> qdisc noqueue master ovs-system state UNKNOWN qlen 1000 >>> link/ether 02:c0:7a:e3:4e:76 brd ff:ff:ff:ff:ff:ff >>> inet6 fe80::c0:7aff:fee3:4e76/64 scope link >>> valid_lft forever preferred_lft forever >>> 24: ;vdsmdummy;: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state >>> DOWN qlen 1000 >>> link/ether a2:2f:f2:58:88:da brd ff:ff:ff:ff:ff:ff >>> 26: vnet0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc >>> pfifo_fast master ovs-system state UNKNOWN qlen 1000 >>> link/ether fe:1a:4a:16:01:03 brd ff:ff:ff:ff:ff:ff >>> inet6 fe80::fc1a:4aff:fe16:103/64 scope link >>> valid_lft forever preferred_lft forever >>> 29: vnet1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc >>> pfifo_fast master ovirtmgmt state UNKNOWN qlen 1000 >>> link/ether fe:1a:4a:16:01:05 brd ff:ff:ff:ff:ff:ff >>> inet6 fe80::fc1a:4aff:fe16:105/64 scope link >>> valid_lft forever preferred_lft forever >>> >>> [root@o2 ~]# ovs-vsctl show >>> 6be6d37c-74cf-485e-9957-f8eb4bddb2ca >>> Bridge br-int >>> fail_mode: secure >>> Port br-int >>> Interface br-int >>> type: internal >>> Port "ovn-972f1b-0" >>> Interface "ovn-972f1b-0" >>> type: geneve >>> options: {csum="true", key=flow, >>> remote_ip="10.0.1.21"} >>> Port "vnet0" >>> Interface "vnet0" >>> ovs_version: "2.9.0" >>> >>> [root@o3 ~]# ip addr >>> 1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN >>> qlen 1 >>> link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 >>> inet 127.0.0.1/8 [1] scope host lo >>> valid_lft forever preferred_lft forever >>> inet6 ::1/128 scope host >>> valid_lft forever preferred_lft forever >>> 2: enp0s31f6: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc >>> pfifo_fast master ovirtmgmt state UP qlen 1000 >>> link/ether 78:f2:9e:90:bc:50 brd ff:ff:ff:ff:ff:ff >>> 3: enp0s20f0u5c2: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc >>> pfifo_fast master public state UNKNOWN qlen 1000 >>> link/ether 50:3e:aa:4c:9c:03 brd ff:ff:ff:ff:ff:ff >>> 4: ovs-system: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN >>> qlen 1000 >>> link/ether 7e:43:c1:b0:48:73 brd ff:ff:ff:ff:ff:ff >>> 5: br-int: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN qlen >>> 1000 >>> link/ether 3a:fe:68:34:31:4c brd ff:ff:ff:ff:ff:ff >>> 21: public: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue >>> state UP qlen 1000 >>> link/ether 50:3e:aa:4c:9c:03 brd ff:ff:ff:ff:ff:ff >>> inet6 fe80::523e:aaff:fe4c:9c03/64 scope link >>> valid_lft forever preferred_lft forever >>> 22: ovirtmgmt: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc >>> noqueue state UP qlen 1000 >>> link/ether 78:f2:9e:90:bc:50 brd ff:ff:ff:ff:ff:ff >>> inet 10.0.1.21/24 [3] brd 10.0.1.255 scope global ovirtmgmt >>> >>> valid_lft forever preferred_lft forever >>> inet6 fe80::7af2:9eff:fe90:bc50/64 scope link >>> valid_lft forever preferred_lft forever >>> 24: ;vdsmdummy;: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state >>> DOWN qlen 1000 >>> link/ether 02:92:3f:89:f2:c7 brd ff:ff:ff:ff:ff:ff >>> 25: vnet0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc >>> pfifo_fast master ovirtmgmt state UNKNOWN qlen 1000 >>> link/ether fe:16:3e:0b:b1:2d brd ff:ff:ff:ff:ff:ff >>> inet6 fe80::fc16:3eff:fe0b:b12d/64 scope link >>> valid_lft forever preferred_lft forever >>> 27: vnet2: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc >>> pfifo_fast master ovirtmgmt state UNKNOWN qlen 1000 >>> link/ether fe:1a:4a:16:01:0b brd ff:ff:ff:ff:ff:ff >>> inet6 fe80::fc1a:4aff:fe16:10b/64 scope link >>> valid_lft forever preferred_lft forever >>> 29: vnet4: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc >>> pfifo_fast master ovs-system state UNKNOWN qlen 1000 >>> link/ether fe:1a:4a:16:01:0c brd ff:ff:ff:ff:ff:ff >>> inet6 fe80::fc1a:4aff:fe16:10c/64 scope link >>> valid_lft forever preferred_lft forever >>> 31: vnet6: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc >>> pfifo_fast master ovirtmgmt state UNKNOWN qlen 1000 >>> link/ether fe:1a:4a:16:01:07 brd ff:ff:ff:ff:ff:ff >>> inet6 fe80::fc1a:4aff:fe16:107/64 scope link >>> valid_lft forever preferred_lft forever >>> 32: vnet7: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc >>> pfifo_fast master public state UNKNOWN qlen 1000 >>> link/ether fe:1a:4a:16:01:09 brd ff:ff:ff:ff:ff:ff >>> inet6 fe80::fc1a:4aff:fe16:109/64 scope link >>> valid_lft forever preferred_lft forever >>> 33: vnet8: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc >>> pfifo_fast master ovs-system state UNKNOWN qlen 1000 >>> link/ether fe:1a:4a:16:01:0a brd ff:ff:ff:ff:ff:ff >>> inet6 fe80::fc1a:4aff:fe16:10a/64 scope link >>> valid_lft forever preferred_lft forever >>> 34: genev_sys_6081: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 65000 >>> qdisc noqueue master ovs-system state UNKNOWN qlen 1000 >>> link/ether 46:88:1c:22:6f:c3 brd ff:ff:ff:ff:ff:ff >>> inet6 fe80::4488:1cff:fe22:6fc3/64 scope link >>> valid_lft forever preferred_lft forever >>> >>> [root@o3 ~]# ovs-vsctl show >>> 8c2c19fc-d9e4-423d-afcb-f5ecff602ca7 >>> Bridge br-int >>> fail_mode: secure >>> Port "vnet4" >>> Interface "vnet4" >>> Port "ovn-049183-0" >>> Interface "ovn-049183-0" >>> type: geneve >>> options: {csum="true", key=flow, >>> remote_ip="10.0.1.18"} >>> Port "vnet8" >>> Interface "vnet8" >>> Port br-int >>> Interface br-int >>> type: internal >>> ovs_version: "2.9.0" >>> >>> Best regards, >>> Samuli >>> >>> Marcin Mirecki wrote: >>> >>> Hi Samuli, >>> >>> Let's first make sure the configuration is correct. >>> How did you configure the env? Did you use the automatic >>> engine-setup >>> configuration? >>> >>> Can you please send me the output of the following: >>> >>> on engine: >>> ovn-sbctl show >>> ovn-nbctl show >>> >>> on hosts: >>> ip addr >>> ovs-vsctl show >>> >>> The 'vdsm-tool ovn-config' command configures the ovn controller to >>> use the >>> first ip as the ovn central, and the local tunnel to use the second >>> one. >>> >>> Regards, >>> Marcin >>> >>> On Sun, May 6, 2018 at 10:42 AM, Samuli >>> Heinonen<[email protected]> >>> wrote: >>> >>> Hi all, >>> >>> I'm building a home lab using oVirt+GlusterFS in hyperconverged(ish) >>> setup. >>> >>> My setup consists of 2x nodes with ASRock H110M-STX motherboard, >>> Intel >>> Pentium G4560 3,5 GHz CPU and 16 GB RAM. Motherboard has integrated >>> Intel >>> Gigabit I219V LAN. At the moment I'm using RaspberryPi as Gluster >>> arbiter >>> node. Nodes are connected to basic "desktop switch" without any >>> management >>> available. >>> >>> Hardware is nowhere near perfect, but it get its job done and is >>> enough >>> for playing around. However I'm having problems getting OVN to work >>> properly and I'm clueless where to look next. >>> >>> oVirt is setup like this: >>> oVirt engine host oe / 10.0.1.101 >>> oVirt hypervisor host o2 / 10.0.1.18 >>> oVirt hypervisor host o3 / 10.0.1.21 >>> OVN network 10.0.200.0/24 [4] >>> >>> When I spin up a VM in o2 and o3 with IP address in network >>> 10.0.1.0/24 [5] >>> >>> everything works fine. VMs can interact between each other without >>> any >>> problems. >>> >>> Problems show up when I try to use OVN based network between virtual >>> machines. If virtual machines are on same hypervisor then everything >>> seems >>> to work ok. But if I have virtual machine on hypervisor o2 and >>> another one >>> on hypervisor o3 then TCP connections doesn't work very well. UDP >>> seems to >>> be ok and it's possible to ping hosts, do dns& ntp queries and so >>> on. >>> >>> Problem with TCP is that for example when taking SSH connection to >>> another >>> host at some point connection just hangs and most of the time it's >>> not even >>> possible to even log in before connectiong hangs. If I look into >>> tcpdump at >>> that point it looks like packets never reach destination. Also, if I >>> have >>> multiple connections, then all of them hang at the same time. >>> >>> I have tried switching off tx checksum and other similar settings, >>> but it >>> didn't make any difference. >>> >>> I'm suspecting that hardware is not good enough. Before >>> investigating into >>> new hardware I'd like to get some confirmation that everything is >>> setup >>> correctly. >>> >>> When setting up oVirt/OVN I had to run following undocumented >>> command to >>> get it working at all: vdsm-tool ovn-config 10.0.1.101 10.0.1.21 >>> (oVirt >>> engine IP, hypervisor IP). Especially this makes me think that I >>> have >>> missed some crucial part in configuration. >>> >>> On oVirt engine in /var/log/openvswitch/ovsdb-server-nb.log there >>> are >>> error messages: >>> 2018-05-06T08:30:05.418Z|00913|stream_ssl|WARN|SSL_read: unexpected >>> SSL >>> connection close >>> 2018-05-06T08:30:05.418Z|00914|jsonrpc|WARN|ssl:127.0.0.1:53152 [6]: >>> receive >>> error: Protocol error >>> 2018-05-06T08:30:05.419Z|00915|reconnect|WARN|ssl:127.0.0.1:53152 >>> [6]: >>> connection dropped (Protocol error) >>> >>> To be honest, I'm not sure what's causing those error messages or >>> are they >>> related. I found out some bug reports stating that they are not >>> critical. >>> >>> Any ideas what to do next or should I just get better hardware? :) >>> >>> Best regards, >>> Samuli Heinonen >>> _______________________________________________ >>> Users mailing list >>> [email protected] >>> http://lists.ovirt.org/mailman/listinfo/users [7] >>> >> >> >> >> Links: >> ------ >> [1] http://127.0.0.1/8 >> [2] http://10.0.1.18/24 >> [3] http://10.0.1.21/24 >> [4] http://10.0.200.0/24 >> [5] http://10.0.1.0/24 >> [6] http://127.0.0.1:53152 >> [7] http://lists.ovirt.org/mailman/listinfo/users >> >
_______________________________________________ Users mailing list -- [email protected] To unsubscribe send an email to [email protected]

