On Thu, Aug 22, 2019 at 3:06 PM Curtis E. Combs Jr. <ej.alb...@gmail.com>
wrote:

> Seems like the STP options are so common and necessary that it would
> be a priority over seldom-used bridge_opts. I know what STP is and I'm
> not even a networking guy - never even heard of half of the
> bridge_opts that have switches in the UI.
>
> Anyway. I wanted to try the openvswitches, so I reinstalled all of my
> nodes and used "openvswitch (Technology Preview)" as the engine-setup
> option for the first host. I made a new Cluster for my nodes, added
> them all to the new cluster, created a new "logical network" for the
> internal network and attached it to the internal network ports.
>
> Now, when I go to create a new VM, I don't even have either the
> ovirtmgmt switch OR the internal switch as an option. The drop-down is
> empy as if I don't have any vnic-profiles.
>
>
openvswitch clusters are limited to ovn networks.
You can create one like described in
https://www.ovirt.org/documentation/admin-guide/chap-External_Providers.html#connecting-an-ovn-network-to-a-physical-network



> On Thu, Aug 22, 2019 at 7:34 AM Tony Pearce <tony...@gmail.com> wrote:
> >
> > Hi Dominik, would you mind sharing the use case for stp via API Only? I
> am keen to know this.
> > Thanks
> >
> >
> > On Thu., 22 Aug. 2019, 19:24 Dominik Holler, <dhol...@redhat.com> wrote:
> >>
> >>
> >>
> >> On Thu, Aug 22, 2019 at 1:08 PM Miguel Duarte de Mora Barroso <
> mdbarr...@redhat.com> wrote:
> >>>
> >>> On Sat, Aug 17, 2019 at 11:27 AM <ej.alb...@gmail.com> wrote:
> >>> >
> >>> > Hello. I have been trying to figure out an issue for a very long
> time.
> >>> > That issue relates to the ethernet and 10gb fc links that I have on
> my
> >>> > cluster being disabled any time a migration occurs.
> >>> >
> >>> > I believe this is because I need to have STP turned on in order to
> >>> > participate with the switch. However, there does not seem to be any
> >>> > way to tell oVirt to stop turning it off! Very frustrating.
> >>> >
> >>> > After entering a cronjob that enables stp on all bridges every 1
> >>> > minute, the migration issue disappears....
> >>> >
> >>> > Is there any way at all to do without this cronjob and set STP to be
> >>> > ON without having to resort to such a silly solution?
> >>>
> >>> Vdsm exposes a per bridge STP knob that you can use for this. By
> >>> default it is set to false, which is probably why you had to use this
> >>> shenanigan.
> >>>
> >>> You can, for instance:
> >>>
> >>> # show present state
> >>> [vagrant@vdsm ~]$ ip a
> >>> 1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN
> >>> group default qlen 1000
> >>>     link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
> >>>     inet 127.0.0.1/8 scope host lo
> >>>        valid_lft forever preferred_lft forever
> >>> 2: eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast
> >>> state UP group default qlen 1000
> >>>     link/ether 52:54:00:41:fb:37 brd ff:ff:ff:ff:ff:ff
> >>> 3: eth1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast
> >>> state UP group default qlen 1000
> >>>     link/ether 52:54:00:83:5b:6f brd ff:ff:ff:ff:ff:ff
> >>>     inet 192.168.50.50/24 brd 192.168.50.255 scope global
> noprefixroute eth1
> >>>        valid_lft forever preferred_lft forever
> >>>     inet6 fe80::5054:ff:fe83:5b6f/64 scope link
> >>>        valid_lft forever preferred_lft forever
> >>> 19: ;vdsmdummy;: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN
> >>> group default qlen 1000
> >>>     link/ether 8e:5c:2e:87:fa:0b brd ff:ff:ff:ff:ff:ff
> >>>
> >>> # show example bridge configuration - you're looking for the STP knob
> here.
> >>> [root@vdsm ~]$ cat bridged_net_with_stp
> >>> {
> >>>   "bondings": {},
> >>>   "networks": {
> >>>     "test-network": {
> >>>       "nic": "eth0",
> >>>       "switch": "legacy",
> >>>       "bridged": true,
> >>>       "stp": true
> >>>     }
> >>>   },
> >>>   "options": {
> >>>     "connectivityCheck": false
> >>>   }
> >>> }
> >>>
> >>> # issue setup networks command:
> >>> [root@vdsm ~]$ vdsm-client -f bridged_net_with_stp Host setupNetworks
> >>> {
> >>>     "code": 0,
> >>>     "message": "Done"
> >>> }
> >>>
> >>> # show bridges
> >>> [root@vdsm ~]$ brctl show
> >>> bridge name bridge id STP enabled interfaces
> >>> ;vdsmdummy; 8000.000000000000 no
> >>> test-network 8000.52540041fb37 yes eth0
> >>>
> >>> # show final state
> >>> [root@vdsm ~]$ ip a
> >>> 1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN
> >>> group default qlen 1000
> >>>     link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
> >>>     inet 127.0.0.1/8 scope host lo
> >>>        valid_lft forever preferred_lft forever
> >>> 2: eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast
> >>> master test-network state UP group default qlen 1000
> >>>     link/ether 52:54:00:41:fb:37 brd ff:ff:ff:ff:ff:ff
> >>> 3: eth1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast
> >>> state UP group default qlen 1000
> >>>     link/ether 52:54:00:83:5b:6f brd ff:ff:ff:ff:ff:ff
> >>>     inet 192.168.50.50/24 brd 192.168.50.255 scope global
> noprefixroute eth1
> >>>        valid_lft forever preferred_lft forever
> >>>     inet6 fe80::5054:ff:fe83:5b6f/64 scope link
> >>>        valid_lft forever preferred_lft forever
> >>> 19: ;vdsmdummy;: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN
> >>> group default qlen 1000
> >>>     link/ether 8e:5c:2e:87:fa:0b brd ff:ff:ff:ff:ff:ff
> >>> 432: test-network: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc
> >>> noqueue state UP group default qlen 1000
> >>>     link/ether 52:54:00:41:fb:37 brd ff:ff:ff:ff:ff:ff
> >>>
> >>> I don't think this STP parameter is exposed via engine UI; @Dominik
> >>> Holler , could you confirm ? What are our plans for it ?
> >>>
> >>
> >> STP is only available via REST-API, see
> >> http://ovirt.github.io/ovirt-engine-api-model/4.3/#types/network
> >> please find an example how to enable STP in
> >> https://gist.github.com/dominikholler/4e70c9ef9929d93b6807f56d43a70b95
> >>
> >> We have no plans to add STP to the web ui,
> >> but new feature requests are always welcome on
> >> https://bugzilla.redhat.com/enter_bug.cgi?product=ovirt-engine
> >>
> >>
> >>>
> >>> >
> >>> > Here are some details about my systems, if you need it.
> >>> >
> >>> >
> >>> > selinux is disabled.
> >>> >
> >>> >
> >>> >
> >>> >
> >>> >
> >>> >
> >>> >
> >>> >
> >>> >
> >>> > [root@swm-02 ~]# rpm -qa | grep ovirt
> >>> > ovirt-imageio-common-1.5.1-0.el7.x86_64
> >>> > ovirt-release43-4.3.5.2-1.el7.noarch
> >>> > ovirt-imageio-daemon-1.5.1-0.el7.noarch
> >>> > ovirt-vmconsole-host-1.0.7-2.el7.noarch
> >>> > ovirt-hosted-engine-setup-2.3.11-1.el7.noarch
> >>> > ovirt-ansible-hosted-engine-setup-1.0.26-1.el7.noarch
> >>> > python2-ovirt-host-deploy-1.8.0-1.el7.noarch
> >>> > ovirt-ansible-engine-setup-1.1.9-1.el7.noarch
> >>> > python2-ovirt-setup-lib-1.2.0-1.el7.noarch
> >>> > cockpit-machines-ovirt-195.1-1.el7.noarch
> >>> > ovirt-hosted-engine-ha-2.3.3-1.el7.noarch
> >>> > ovirt-vmconsole-1.0.7-2.el7.noarch
> >>> > cockpit-ovirt-dashboard-0.13.5-1.el7.noarch
> >>> > ovirt-provider-ovn-driver-1.2.22-1.el7.noarch
> >>> > ovirt-host-deploy-common-1.8.0-1.el7.noarch
> >>> > ovirt-host-4.3.4-1.el7.x86_64
> >>> > python-ovirt-engine-sdk4-4.3.2-2.el7.x86_64
> >>> > ovirt-host-dependencies-4.3.4-1.el7.x86_64
> >>> > ovirt-ansible-repositories-1.1.5-1.el7.noarch
> >>> > [root@swm-02 ~]# cat /etc/redhat-release
> >>> > CentOS Linux release 7.6.1810 (Core)
> >>> > [root@swm-02 ~]# uname -r
> >>> > 3.10.0-957.27.2.el7.x86_64
> >>> > You have new mail in /var/spool/mail/root
> >>> > [root@swm-02 ~]# ip a
> >>> > 1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN
> >>> > group default qlen 1000
> >>> >     link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
> >>> >     inet 127.0.0.1/8 scope host lo
> >>> >        valid_lft forever preferred_lft forever
> >>> > 2: em1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq master
> >>> > test state UP group default qlen 1000
> >>> >     link/ether d4:ae:52:8d:50:48 brd ff:ff:ff:ff:ff:ff
> >>> > 3: em2: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN group
> >>> > default qlen 1000
> >>> >     link/ether d4:ae:52:8d:50:49 brd ff:ff:ff:ff:ff:ff
> >>> > 4: p1p1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq master
> >>> > ovirtmgmt state UP group default qlen 1000
> >>> >     link/ether 90:e2:ba:1e:14:80 brd ff:ff:ff:ff:ff:ff
> >>> > 5: p1p2: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN group
> >>> > default qlen 1000
> >>> >     link/ether 90:e2:ba:1e:14:81 brd ff:ff:ff:ff:ff:ff
> >>> > 6: ovs-system: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN
> >>> > group default qlen 1000
> >>> >     link/ether a2:b8:d6:e8:b3:d8 brd ff:ff:ff:ff:ff:ff
> >>> > 7: br-int: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN group
> >>> > default qlen 1000
> >>> >     link/ether 96:a0:c1:4a:45:4b brd ff:ff:ff:ff:ff:ff
> >>> > 25: test: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue
> >>> > state UP group default qlen 1000
> >>> >     link/ether d4:ae:52:8d:50:48 brd ff:ff:ff:ff:ff:ff
> >>> >     inet 10.15.11.21/24 brd 10.15.11.255 scope global test
> >>> >        valid_lft forever preferred_lft forever
> >>> > 26: ovirtmgmt: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc
> >>> > noqueue state UP group default qlen 1000
> >>> >     link/ether 90:e2:ba:1e:14:80 brd ff:ff:ff:ff:ff:ff
> >>> >     inet 10.15.28.31/24 brd 10.15.28.255 scope global ovirtmgmt
> >>> >        valid_lft forever preferred_lft forever
> >>> > 27: ;vdsmdummy;: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN
> >>> > group default qlen 1000
> >>> >     link/ether 62:e5:e5:07:99:eb brd ff:ff:ff:ff:ff:ff
> >>> > 29: vnet0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq master
> >>> > ovirtmgmt state UNKNOWN group default qlen 1000
> >>> >     link/ether fe:6f:9c:95:00:02 brd ff:ff:ff:ff:ff:ff
> >>> > [root@swm-02 ~]# free -m
> >>> >               total        used        free      shared  buff/cache
>  available
> >>> > Mem:          64413        1873       61804           9         735
>      62062
> >>> > Swap:         16383           0       16383
> >>> > [root@swm-02 ~]# free -h
> >>> >               total        used        free      shared  buff/cache
>  available
> >>> > Mem:            62G        1.8G         60G        9.5M        735M
>        60G
> >>> > Swap:           15G          0B         15G
> >>> > [root@swm-02 ~]# ls
> >>> > ls                  lsb_release         lshw                lslocks
> >>> >          lsmod               lspci               lssubsys
> >>> > lsusb.py
> >>> > lsattr              lscgroup            lsinitrd            lslogins
> >>> >          lsns                lss16toppm          lstopo-no-graphics
> >>> > lsblk               lscpu               lsipc               lsmem
> >>> >          lsof                lsscsi              lsusb
> >>> > [root@swm-02 ~]# lscpu
> >>> > Architecture:          x86_64
> >>> > CPU op-mode(s):        32-bit, 64-bit
> >>> > Byte Order:            Little Endian
> >>> > CPU(s):                16
> >>> > On-line CPU(s) list:   0-15
> >>> > Thread(s) per core:    2
> >>> > Core(s) per socket:    4
> >>> > Socket(s):             2
> >>> > NUMA node(s):          2
> >>> > Vendor ID:             GenuineIntel
> >>> > CPU family:            6
> >>> > Model:                 44
> >>> > Model name:            Intel(R) Xeon(R) CPU           X5672  @
> 3.20GHz
> >>> > Stepping:              2
> >>> > CPU MHz:               3192.064
> >>> > BogoMIPS:              6384.12
> >>> > Virtualization:        VT-x
> >>> > L1d cache:             32K
> >>> > L1i cache:             32K
> >>> > L2 cache:              256K
> >>> > L3 cache:              12288K
> >>> > NUMA node0 CPU(s):     0,2,4,6,8,10,12,14
> >>> > NUMA node1 CPU(s):     1,3,5,7,9,11,13,15
> >>> > Flags:                 fpu vme de pse tsc msr pae mce cx8 apic sep
> >>> > mtrr pge mca cmov pat pse36 clflush dts acpi mmx fxsr sse sse2 ss ht
> >>> > tm pbe syscall nx pdpe1gb rdtscp lm constant_tsc arch_perfmon pebs
> bts
> >>> > rep_good nopl xtopology nonstop_tsc aperfmperf eagerfpu pni pclmulqdq
> >>> > dtes64 monitor ds_cpl vmx smx est tm2 ssse3 cx16 xtpr pdcm pcid dca
> >>> > sse4_1 sse4_2 popcnt aes lahf_lm ssbd ibrs ibpb stibp tpr_shadow vnmi
> >>> > flexpriority ept vpid dtherm ida arat spec_ctrl intel_stibp flush_l1d
> >>> > [root@swm-02 ~]#
> >>> > _______________________________________________
> >>> > Users mailing list -- users@ovirt.org
> >>> > To unsubscribe send an email to users-le...@ovirt.org
> >>> > Privacy Statement: https://www.ovirt.org/site/privacy-policy/
> >>> > oVirt Code of Conduct:
> https://www.ovirt.org/community/about/community-guidelines/
> >>> > List Archives:
> https://lists.ovirt.org/archives/list/users@ovirt.org/message/MTMZ5MF4CF2VR2D25VVPDNFN2IKE24AR/
> >>
> >> _______________________________________________
> >> Users mailing list -- users@ovirt.org
> >> To unsubscribe send an email to users-le...@ovirt.org
> >> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
> >> oVirt Code of Conduct:
> https://www.ovirt.org/community/about/community-guidelines/
> >> List Archives:
> https://lists.ovirt.org/archives/list/users@ovirt.org/message/QBA7NYKAJNREIV6TN42VCW4IN3CX4VFG/
>
_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/LFLKQJCPJK5FXH46PCWNKTXMQN73HTLX/

Reply via email to