Please push the patch into: https://gerrit.ovirt.org/ovirt-provider-ovn (let me know if you need some directions)
----- Original Message ----- > From: "Sverker Abrahamsson" <sver...@abrahamsson.com> > To: "Marcin Mirecki" <mmire...@redhat.com> > Cc: "Ovirt Users" <users@ovirt.org> > Sent: Monday, January 9, 2017 1:45:37 PM > Subject: Re: [ovirt-users] Issue with OVN/OVS and mandatory ovirtmgmt network > > Ok, found it. The issue is right here: > > <interface type="bridge"> > <mac address="00:1a:4a:16:01:54" /> > <model type="virtio" /> > <source bridge="br-int" /> > <virtualport type="openvswitch" /> > <link state="up" /> > <boot order="2" /> > <bandwidth /> > <virtualport type="openvswitch"> > <parameters > interfaceid="912cba79-982e-4a87-868e-241fedccb59a" /> > </virtualport> > </interface> > > There are two elements for virtualport, the first without id and the > second with. On h2 I had fixed this which was the patch I posted earlier > although I switched back to use br-int after understanding that was the > correct way. When that hook was copied to h1 the port gets attached fine. > > Patch with updated testcase attached. > > /Sverker > > > Den 2017-01-09 kl. 10:41, skrev Sverker Abrahamsson: > > This is the content of vdsm.log on h1 at this time: > > > > 2017-01-06 20:54:12,636 INFO (jsonrpc/7) [jsonrpc.JsonRpcServer] RPC > > call VM.create succeeded in 0.01 seconds (__init__:515) > > 2017-01-06 20:54:12,636 INFO (vm/6dd5291e) [virt.vm] > > (vmId='6dd5291e-6556-4d29-8b4e-ea896e627645') VM wrapper has started > > (vm:1901) > > 2017-01-06 20:54:12,636 INFO (vm/6dd5291e) [vds] prepared volume > > path: > > /rhev/data-center/mnt/h2-int.limetransit.com:_var_lib_exports_iso/1d49c4bc-0fec-4503-a583-d476fa3a370d/images/11111111-1111-1111-1111-111111111111/CentOS-7-x86_64-NetInstall-1611.iso > > (clientIF:374) > > 2017-01-06 20:54:12,743 INFO (vm/6dd5291e) [root] (hooks:108) > > 2017-01-06 20:54:12,847 INFO (vm/6dd5291e) [root] (hooks:108) > > 2017-01-06 20:54:12,863 INFO (vm/6dd5291e) [virt.vm] > > (vmId='6dd5291e-6556-4d29-8b4e-ea896e627645') <?xml version='1.0' > > encoding='UTF-8'?> > > <domain xmlns:ovirt="http://ovirt.org/vm/tune/1.0" type="kvm"> > > <name>CentOS7_3</name> > > <uuid>6dd5291e-6556-4d29-8b4e-ea896e627645</uuid> > > <memory>1048576</memory> > > <currentMemory>1048576</currentMemory> > > <maxMemory slots="16">4294967296</maxMemory> > > <vcpu current="1">16</vcpu> > > <devices> > > <channel type="unix"> > > <target name="com.redhat.rhevm.vdsm" type="virtio" /> > > <source mode="bind" > > path="/var/lib/libvirt/qemu/channels/6dd5291e-6556-4d29-8b4e-ea896e627645.com.redhat.rhevm.vdsm" > > /> > > </channel> > > <channel type="unix"> > > <target name="org.qemu.guest_agent.0" type="virtio" /> > > <source mode="bind" > > path="/var/lib/libvirt/qemu/channels/6dd5291e-6556-4d29-8b4e-ea896e627645.org.qemu.guest_agent.0" > > /> > > </channel> > > <input bus="ps2" type="mouse" /> > > <memballoon model="virtio" /> > > <controller index="0" model="virtio-scsi" type="scsi" /> > > <controller index="0" ports="16" type="virtio-serial" /> > > <video> > > <model heads="1" ram="65536" type="qxl" vgamem="16384" > > vram="32768" /> > > </video> > > <graphics autoport="yes" defaultMode="secure" passwd="*****" > > passwdValidTo="1970-01-01T00:00:01" port="-1" tlsPort="-1" type="spice"> > > <channel mode="secure" name="main" /> > > <channel mode="secure" name="inputs" /> > > <channel mode="secure" name="cursor" /> > > <channel mode="secure" name="playback" /> > > <channel mode="secure" name="record" /> > > <channel mode="secure" name="display" /> > > <channel mode="secure" name="smartcard" /> > > <channel mode="secure" name="usbredir" /> > > <listen network="vdsm-ovirtmgmt" type="network" /> > > </graphics> > > <interface type="bridge"> > > <mac address="00:1a:4a:16:01:54" /> > > <model type="virtio" /> > > <source bridge="br-int" /> > > <virtualport type="openvswitch" /> > > <link state="up" /> > > <boot order="2" /> > > <bandwidth /> > > <virtualport type="openvswitch"> > > <parameters > > interfaceid="912cba79-982e-4a87-868e-241fedccb59a" /> > > </virtualport> > > </interface> > > <disk device="cdrom" snapshot="no" type="file"> > > <source > > file="/rhev/data-center/mnt/h2-int.limetransit.com:_var_lib_exports_iso/1d49c4bc-0fec-4503-a583-d476fa3a370d/images/11111111-1111-1111-1111-111111111111/CentOS-7-x86_64-NetInstall-1611.iso" > > startupPolicy="optional" /> > > <target bus="ide" dev="hdc" /> > > <readonly /> > > <boot order="1" /> > > </disk> > > <channel type="spicevmc"> > > <target name="com.redhat.spice.0" type="virtio" /> > > </channel> > > </devices> > > <metadata> > > <ovirt:qos /> > > </metadata> > > <os> > > <type arch="x86_64" machine="pc-i440fx-rhel7.2.0">hvm</type> > > <smbios mode="sysinfo" /> > > <bootmenu enable="yes" timeout="10000" /> > > </os> > > <sysinfo type="smbios"> > > <system> > > <entry name="manufacturer">oVirt</entry> > > <entry name="product">oVirt Node</entry> > > <entry name="version">7-3.1611.el7.centos</entry> > > <entry > > name="serial">62f1adff-b29e-4a7c-abba-c2c4c73248c6</entry> > > <entry > > name="uuid">6dd5291e-6556-4d29-8b4e-ea896e627645</entry> > > </system> > > </sysinfo> > > <clock adjustment="0" offset="variable"> > > <timer name="rtc" tickpolicy="catchup" /> > > <timer name="pit" tickpolicy="delay" /> > > <timer name="hpet" present="no" /> > > </clock> > > <features> > > <acpi /> > > </features> > > <cpu match="exact"> > > <model>SandyBridge</model> > > <topology cores="1" sockets="16" threads="1" /> > > <numa> > > <cell cpus="0" memory="1048576" /> > > </numa> > > </cpu> > > </domain> > > (vm:1988) > > 2017-01-06 20:54:13,046 INFO (libvirt/events) [virt.vm] > > (vmId='6dd5291e-6556-4d29-8b4e-ea896e627645') CPU running: onResume > > (vm:4863) > > 2017-01-06 20:54:13,058 INFO (vm/6dd5291e) [virt.vm] > > (vmId='6dd5291e-6556-4d29-8b4e-ea896e627645') Starting connection > > (guestagent:245) > > 2017-01-06 20:54:13,060 INFO (vm/6dd5291e) [virt.vm] > > (vmId='6dd5291e-6556-4d29-8b4e-ea896e627645') CPU running: domain > > initialization (vm:4863) > > 2017-01-06 20:54:15,154 INFO (jsonrpc/6) [jsonrpc.JsonRpcServer] RPC > > call Host.getVMFullList succeeded in 0.01 seconds (__init__:515) > > 2017-01-06 20:54:17,571 INFO (periodic/2) [dispatcher] Run and > > protect: getVolumeSize(sdUUID=u'2ee54fb8-48f2-4576-8cff-f2346504b08b', > > spUUID=u'584ebd64-0268-0193-025b-00000000038e', > > imgUUID=u'5a3aae57-ffe0-4a3b-aa87-8461669db7f9', > > volUUID=u'b6a88789-fcb1-4d3e-911b-2a4d3b6c69c7', options=None) > > (logUtils:49) > > 2017-01-06 20:54:17,573 INFO (periodic/2) [dispatcher] Run and > > protect: getVolumeSize, Return response: {'truesize': '1859723264', > > 'apparentsize': '21474836480'} (logUtils:52) > > 2017-01-06 20:54:21,211 INFO (periodic/2) [dispatcher] Run and > > protect: repoStats(options=None) (logUtils:49) > > 2017-01-06 20:54:21,212 INFO (periodic/2) [dispatcher] Run and > > protect: repoStats, Return response: > > {u'2ee54fb8-48f2-4576-8cff-f2346504b08b': {'code': 0, 'actual': True, > > 'version': 3, 'acquired': True, 'delay': '0.000936552', 'lastCheck': > > '1.4', 'valid': True}, u'1d49c4bc-0fec-4503-a583-d476fa3a370d': > > {'code': 0, 'actual': True, 'version': 0, 'acquired': True, 'delay': > > '0.000960248', 'lastCheck': '1.4', 'valid': True}} (logUtils:52) > > 2017-01-06 20:54:23,543 INFO (jsonrpc/2) [jsonrpc.JsonRpcServer] RPC > > call Host.getAllVmStats succeeded in 0.00 seconds (__init__:515) > > 2017-01-06 20:54:23,641 INFO (jsonrpc/1) [jsonrpc.JsonRpcServer] RPC > > call Host.getAllVmIoTunePolicies succeeded in 0.00 seconds (__init__:515) > > 2017-01-06 20:54:24,918 INFO (jsonrpc/0) [dispatcher] Run and > > protect: repoStats(options=None) (logUtils:49) > > 2017-01-06 20:54:24,918 INFO (jsonrpc/0) [dispatcher] Run and > > protect: repoStats, Return response: > > {u'2ee54fb8-48f2-4576-8cff-f2346504b08b': {'code': 0, 'actual': True, > > 'version': 3, 'acquired': True, 'delay': '0.000936552', 'lastCheck': > > '5.1', 'valid': True}, u'1d49c4bc-0fec-4503-a583-d476fa3a370d': > > {'code': 0, 'actual': True, 'version': 0, 'acquired': True, 'delay': > > '0.000960248', 'lastCheck': '2.1', 'valid': True}} (logUtils:52) > > 2017-01-06 20:54:24,924 INFO (jsonrpc/0) [jsonrpc.JsonRpcServer] RPC > > call Host.getStats succeeded in 0.01 seconds (__init__:515) > > > > Vdsm and the OVN driver must have been called as the port IS created, > > but with the wrong id. I don't find the faulty id in vdsm.log neither, > > the xml above have the correct id. > > /Sverker > > > > Den 2017-01-09 kl. 10:06, skrev Marcin Mirecki: > >> The port is set up on the host by the ovirt-provider-ovn-driver. > >> The driver is invoked by the vdsm hook whenever any operation on > >> the port is done. > >> Please ensure that this is installed properly. > >> You can check the vdsm log (/var/log/vdsm/vdsm.log) to see if the > >> hook was executed properly. > >> > >> > >> ----- Original Message ----- > >>> From: "Sverker Abrahamsson" <sver...@abrahamsson.com> > >>> To: "Marcin Mirecki" <mmire...@redhat.com> > >>> Cc: "Ovirt Users" <users@ovirt.org> > >>> Sent: Friday, January 6, 2017 9:00:26 PM > >>> Subject: Re: [ovirt-users] Issue with OVN/OVS and mandatory > >>> ovirtmgmt network > >>> > >>> I created a new VM in the ui and assigned it to host h1. In > >>> /var/log/ovirt-provider-ovn.log I get the following: > >>> > >>> 2017-01-06 20:54:11,940 Request: GET : /v2.0/ports > >>> 2017-01-06 20:54:11,940 Connecting to remote ovn database: > >>> tcp:127.0.0.1:6641 > >>> 2017-01-06 20:54:12,157 Connected (number of retries: 2) > >>> 2017-01-06 20:54:12,158 Response code: 200 > >>> 2017-01-06 20:54:12,158 Response body: {"ports": [{"name": > >>> "4981ee5f-6e15-4bd5-a1cf-7ead9bdd5873", "network_id": > >>> "e53554cf-e553-40a1-8d22-9c8d95ec0601", "device_owner": "oVirt", > >>> "mac_address": "00:1a:4a:16:01:51", "id": > >>> "4981ee5f-6e15-4bd5-a1cf-7ead9bdd5873", "device_id": > >>> "40cd7328-d575-4c3d-b656-9ef9bacc0078"}, {"name": > >>> "92f6d3c8-68b3-4986-9c09-60bee04644b5", "network_id": > >>> "e53554cf-e553-40a1-8d22-9c8d95ec0601", "device_owner": "oVirt", > >>> "mac_address": "00:1a:4a:16:01:52", "id": > >>> "92f6d3c8-68b3-4986-9c09-60bee04644b5", "device_id": > >>> "4baefa8c-3822-4de0-9cd0-1d025bab7844"}]} > >>> 2017-01-06 20:54:12,160 Request: SHOW : > >>> /v2.0/networks/e53554cf-e553-40a1-8d22-9c8d95ec0601 > >>> 2017-01-06 20:54:12,160 Connecting to remote ovn database: > >>> tcp:127.0.0.1:6641 > >>> 2017-01-06 20:54:12,377 Connected (number of retries: 2) > >>> 2017-01-06 20:54:12,378 Response code: 200 > >>> 2017-01-06 20:54:12,378 Response body: {"network": {"id": > >>> "e53554cf-e553-40a1-8d22-9c8d95ec0601", "name": "ovirtbridge"}} > >>> 2017-01-06 20:54:12,380 Request: POST : /v2.0/ports > >>> 2017-01-06 20:54:12,380 Request body: > >>> { > >>> "port" : { > >>> "name" : "nic1", > >>> "binding:host_id" : "h1.limetransit.com", > >>> "admin_state_up" : true, > >>> "device_id" : "e8553a88-05f0-401d-8b9b-5fff77f7bbbe", > >>> "device_owner" : "oVirt", > >>> "mac_address" : "00:1a:4a:16:01:54", > >>> "network_id" : "e53554cf-e553-40a1-8d22-9c8d95ec0601" > >>> } > >>> } > >>> 2017-01-06 20:54:12,380 Connecting to remote ovn database: > >>> tcp:127.0.0.1:6641 > >>> 2017-01-06 20:54:12,610 Connected (number of retries: 2) > >>> 2017-01-06 20:54:12,614 Response code: 200 > >>> 2017-01-06 20:54:12,614 Response body: {"port": {"name": > >>> "912cba79-982e-4a87-868e-241fedccb59a", "network_id": > >>> "e53554cf-e553-40a1-8d22-9c8d95ec0601", "device_owner": "oVirt", > >>> "mac_address": "00:1a:4a:16:01:54", "id": > >>> "912cba79-982e-4a87-868e-241fedccb59a", "device_id": > >>> "e8553a88-05f0-401d-8b9b-5fff77f7bbbe"}} > >>> > >>> h1:/var/log/messages > >>> Jan 6 20:54:12 h1 ovs-vsctl: ovs|00001|vsctl|INFO|Called as ovs-vsctl > >>> --timeout=5 -- --if-exists del-port vnet1 -- add-port br-int vnet1 -- > >>> set Interface vnet1 > >>> "external-ids:attached-mac=\"00:1a:4a:16:01:54\"" -- > >>> set Interface vnet1 > >>> "external-ids:iface-id=\"20388407-0f76-41d8-97aa-8e2b5978f908\"" -- set > >>> Interface vnet1 > >>> "external-ids:vm-id=\"6dd5291e-6556-4d29-8b4e-ea896e627645\"" -- set > >>> Interface vnet1 external-ids:iface-status=active > >>> > >>> [root@h2 ~]# ovn-nbctl show > >>> switch e53554cf-e553-40a1-8d22-9c8d95ec0601 (ovirtbridge) > >>> port 4981ee5f-6e15-4bd5-a1cf-7ead9bdd5873 > >>> addresses: ["00:1a:4a:16:01:51"] > >>> port 912cba79-982e-4a87-868e-241fedccb59a > >>> addresses: ["00:1a:4a:16:01:54"] > >>> port 92f6d3c8-68b3-4986-9c09-60bee04644b5 > >>> addresses: ["00:1a:4a:16:01:52"] > >>> port ovirtbridge-port2 > >>> addresses: ["unknown"] > >>> port ovirtbridge-port1 > >>> addresses: ["unknown"] > >>> [root@h2 ~]# ovn-sbctl show > >>> Chassis "6e4dd29f-7607-48d7-8e5a-eef4c6aeefb5" > >>> hostname: "h2.limetransit.com" > >>> Encap geneve > >>> ip: "148.251.126.50" > >>> options: {csum="true"} > >>> Port_Binding "4981ee5f-6e15-4bd5-a1cf-7ead9bdd5873" > >>> Port_Binding "ovirtbridge-port1" > >>> Chassis "4f10fb04-8fb2-48d7-8a3f-ea6444c02cf9" > >>> hostname: "h1.limetransit.com" > >>> Encap geneve > >>> ip: "144.76.84.73" > >>> options: {csum="true"} > >>> Port_Binding "ovirtbridge-port2" > >>> Port_Binding "92f6d3c8-68b3-4986-9c09-60bee04644b5" > >>> > >>> I.e. same issue > >>> /Sverker > >>> > >>> Den 2017-01-06 kl. 20:49, skrev Sverker Abrahamsson: > >>>> The port is created from Ovirt UI, the ovs-vsctl command below is > >>>> executed when VM is started. In /var/log/ovirt-provider-ovn.log on h2 > >>>> I get the following: > >>>> > >>>> 2017-01-06 20:19:25,452 Request: GET : /v2.0/ports > >>>> 2017-01-06 20:19:25,452 Connecting to remote ovn database: > >>>> tcp:127.0.0.1:6641 > >>>> 2017-01-06 20:19:25,670 Connected (number of retries: 2) > >>>> 2017-01-06 20:19:25,670 Response code: 200 > >>>> 2017-01-06 20:19:25,670 Response body: {"ports": [{"name": > >>>> "4981ee5f-6e15-4bd5-a1cf-7ead9bdd5873", "network_id": > >>>> "e53554cf-e553-40a1-8d22-9c8d95ec0601", "device_owner": "oVirt", > >>>> "mac_address": "00:1a:4a:16:01:51", "id": > >>>> "4981ee5f-6e15-4bd5-a1cf-7ead9bdd5873", "device_id": > >>>> "40cd7328-d575-4c3d-b656-9ef9bacc0078"}, {"name": > >>>> "92f6d3c8-68b3-4986-9c09-60bee04644b5", "network_id": > >>>> "e53554cf-e553-40a1-8d22-9c8d95ec0601", "device_owner": "oVirt", > >>>> "mac_address": "00:1a:4a:16:01:52", "id": > >>>> "92f6d3c8-68b3-4986-9c09-60bee04644b5", "device_id": > >>>> "4baefa8c-3822-4de0-9cd0-1d025bab7844"}]} > >>>> 2017-01-06 20:19:25,673 Request: PUT : > >>>> /v2.0/ports/92f6d3c8-68b3-4986-9c09-60bee04644b5 > >>>> 2017-01-06 20:19:25,673 Request body: > >>>> { > >>>> "port" : { > >>>> "binding:host_id" : "h1.limetransit.com", > >>>> "security_groups" : null > >>>> } > >>>> } > >>>> 2017-01-06 20:19:25,673 Connecting to remote ovn database: > >>>> tcp:127.0.0.1:6641 > >>>> 2017-01-06 20:19:25,890 Connected (number of retries: 2) > >>>> 2017-01-06 20:19:25,891 Response code: 200 > >>>> 2017-01-06 20:19:25,891 Response body: {"port": {"name": > >>>> "92f6d3c8-68b3-4986-9c09-60bee04644b5", "network_id": > >>>> "e53554cf-e553-40a1-8d22-9c8d95ec0601", "device_owner": "oVirt", > >>>> "mac_address": "00:1a:4a:16:01:52", "id": > >>>> "92f6d3c8-68b3-4986-9c09-60bee04644b5", "device_id": > >>>> "4baefa8c-3822-4de0-9cd0-1d025bab7844"}} > >>>> > >>>> In /var/log/messages on h1 I get the following: > >>>> > >>>> Jan 6 20:18:56 h1 dbus-daemon: dbus[1339]: [system] Successfully > >>>> activated service 'org.freedesktop.problems' > >>>> Jan 6 20:19:26 h1 ovs-vsctl: ovs|00001|vsctl|INFO|Called as ovs-vsctl > >>>> --timeout=5 -- --if-exists del-port vnet0 -- add-port br-int vnet0 -- > >>>> set Interface vnet0 "external-ids:attached-mac=\"00:1a:4a:16:01:52\"" > >>>> -- set Interface vnet0 > >>>> "external-ids:iface-id=\"72dafda5-03c2-4bb6-bcb6-241fa5c0a1f3\"" -- > >>>> set Interface vnet0 > >>>> "external-ids:vm-id=\"4d0c134a-11a0-40f4-b2fb-c13c17c7251c\"" -- set > >>>> Interface vnet0 external-ids:iface-status=active > >>>> Jan 6 20:19:26 h1 kernel: device vnet0 entered promiscuous mode > >>>> Jan 6 20:19:26 h1 avahi-daemon[1391]: Registering new address record > >>>> for fe80::fc1a:4aff:fe16:152 on vnet0.*. > >>>> Jan 6 20:19:26 h1 systemd-machined: New machine qemu-4-CentOS72. > >>>> Jan 6 20:19:26 h1 systemd: Started Virtual Machine qemu-4-CentOS72. > >>>> Jan 6 20:19:26 h1 systemd: Starting Virtual Machine qemu-4-CentOS72. > >>>> > >>>> [root@h2 ~]# ovn-nbctl show > >>>> switch e53554cf-e553-40a1-8d22-9c8d95ec0601 (ovirtbridge) > >>>> port 4981ee5f-6e15-4bd5-a1cf-7ead9bdd5873 > >>>> addresses: ["00:1a:4a:16:01:51"] > >>>> port 92f6d3c8-68b3-4986-9c09-60bee04644b5 > >>>> addresses: ["00:1a:4a:16:01:52"] > >>>> port ovirtbridge-port2 > >>>> addresses: ["unknown"] > >>>> port ovirtbridge-port1 > >>>> addresses: ["unknown"] > >>>> [root@h2 ~]# ovn-sbctl show > >>>> Chassis "6e4dd29f-7607-48d7-8e5a-eef4c6aeefb5" > >>>> hostname: "h2.limetransit.com" > >>>> Encap geneve > >>>> ip: "148.251.126.50" > >>>> options: {csum="true"} > >>>> Port_Binding "4981ee5f-6e15-4bd5-a1cf-7ead9bdd5873" > >>>> Port_Binding "ovirtbridge-port1" > >>>> Chassis "4f10fb04-8fb2-48d7-8a3f-ea6444c02cf9" > >>>> hostname: "h1.limetransit.com" > >>>> Encap geneve > >>>> ip: "144.76.84.73" > >>>> options: {csum="true"} > >>>> Port_Binding "ovirtbridge-port2" > >>>> > >>>> I.e. the port is set up with the wrong ID and not attached to OVN. > >>>> > >>>> If I correct external-ids:iface-id like this: > >>>> [root@h1 ~]# ovs-vsctl set Interface vnet0 > >>>> "external-ids:iface-id=\"92f6d3c8-68b3-4986-9c09-60bee04644b5\"" > >>>> > >>>> then sb is correct: > >>>> [root@h2 ~]# ovn-sbctl show > >>>> Chassis "6e4dd29f-7607-48d7-8e5a-eef4c6aeefb5" > >>>> hostname: "h2.limetransit.com" > >>>> Encap geneve > >>>> ip: "148.251.126.50" > >>>> options: {csum="true"} > >>>> Port_Binding "4981ee5f-6e15-4bd5-a1cf-7ead9bdd5873" > >>>> Port_Binding "ovirtbridge-port1" > >>>> Chassis "4f10fb04-8fb2-48d7-8a3f-ea6444c02cf9" > >>>> hostname: "h1.limetransit.com" > >>>> Encap geneve > >>>> ip: "144.76.84.73" > >>>> options: {csum="true"} > >>>> Port_Binding "ovirtbridge-port2" > >>>> Port_Binding "92f6d3c8-68b3-4986-9c09-60bee04644b5" > >>>> > >>>> I don't know from where the ID 72dafda5-03c2-4bb6-bcb6-241fa5c0a1f3 > >>>> comes from, doesn't show in any log other than /var/log/messages. > >>>> > >>>> If I do the same exercise on the same host as engine is running on > >>>> then the port for the VM gets the right id and is working from > >>>> beginning. > >>>> /Sverker > >>>> > >>>> Den 2017-01-03 kl. 10:23, skrev Marcin Mirecki: > >>>>> How did you create this port? > >>>>> From the oVirt engine UI? > >>>>> The OVN provider creates the port when you add the port in the > >>>>> engine UI, > >>>>> it is then plugged into the ovs bridge by the VIF driver. > >>>>> Please attach /var/log/ovirt-provider-ovn.log > >>>>> > >>>>> > >>>>> > >>>>> ----- Original Message ----- > >>>>>> From: "Sverker Abrahamsson"<sver...@abrahamsson.com> > >>>>>> To: "Marcin Mirecki"<mmire...@redhat.com> > >>>>>> Cc: "Ovirt Users"<users@ovirt.org> > >>>>>> Sent: Tuesday, January 3, 2017 2:06:22 AM > >>>>>> Subject: Re: [ovirt-users] Issue with OVN/OVS and mandatory > >>>>>> ovirtmgmt > >>>>>> network > >>>>>> > >>>>>> Found an issue with Ovirt - OVN integration. > >>>>>> > >>>>>> Engine and OVN central db running on host h2. Created VM to run > >>>>>> on host > >>>>>> h1, which is started. Ovn db state: > >>>>>> > >>>>>> [root@h2 env3]# ovn-nbctl show > >>>>>> switch e53554cf-e553-40a1-8d22-9c8d95ec0601 (ovirtbridge) > >>>>>> port 4981ee5f-6e15-4bd5-a1cf-7ead9bdd5873 > >>>>>> addresses: ["00:1a:4a:16:01:51"] > >>>>>> port 92f6d3c8-68b3-4986-9c09-60bee04644b5 > >>>>>> addresses: ["00:1a:4a:16:01:52"] > >>>>>> port ovirtbridge-port2 > >>>>>> addresses: ["unknown"] > >>>>>> port ovirtbridge-port1 > >>>>>> addresses: ["unknown"] > >>>>>> [root@h2 env3]# ovn-sbctl show > >>>>>> Chassis "6e4dd29f-7607-48d7-8e5a-eef4c6aeefb5" > >>>>>> hostname: "h2.limetransit.com" > >>>>>> Encap geneve > >>>>>> ip: "148.251.126.50" > >>>>>> options: {csum="true"} > >>>>>> Port_Binding "4981ee5f-6e15-4bd5-a1cf-7ead9bdd5873" > >>>>>> Port_Binding "ovirtbridge-port1" > >>>>>> Chassis "4f10fb04-8fb2-48d7-8a3f-ea6444c02cf9" > >>>>>> hostname: "h1.limetransit.com" > >>>>>> Encap geneve > >>>>>> ip: "144.76.84.73" > >>>>>> options: {csum="true"} > >>>>>> Port_Binding "ovirtbridge-port2" > >>>>>> > >>>>>> Port 92f6d3c8-68b3-4986-9c09-60bee04644b5 is for the new VM which is > >>>>>> started on h1, but it is not assigned to that chassis. The reason is > >>>>>> that on h1 the port on br-int is created like this: > >>>>>> > >>>>>> ovs-vsctl --timeout=5 -- --if-exists del-port vnet0 -- add-port > >>>>>> br-int > >>>>>> vnet0 -- set Interface vnet0 > >>>>>> "external-ids:attached-mac=\"00:1a:4a:16:01:52\"" -- set > >>>>>> Interface vnet0 > >>>>>> "external-ids:iface-id=\"35bcbe31-2c7e-4d97-add9-ce150eeb2f11\"" > >>>>>> -- set > >>>>>> Interface vnet0 > >>>>>> "external-ids:vm-id=\"4d0c134a-11a0-40f4-b2fb-c13c17c7251c\"" -- set > >>>>>> Interface vnet0 external-ids:iface-status=active > >>>>>> > >>>>>> I.e. the extrernal id of interface is wrong. When I manually > >>>>>> change to > >>>>>> the right id like this the port works fine: > >>>>>> > >>>>>> ovs-vsctl --timeout=5 -- --if-exists del-port vnet0 -- add-port > >>>>>> br-int > >>>>>> vnet0 -- set Interface vnet0 > >>>>>> "external-ids:attached-mac=\"00:1a:4a:16:01:52\"" -- set > >>>>>> Interface vnet0 > >>>>>> "external-ids:iface-id=\"92f6d3c8-68b3-4986-9c09-60bee04644b5\"" > >>>>>> -- set > >>>>>> Interface vnet0 > >>>>>> "external-ids:vm-id=\"4d0c134a-11a0-40f4-b2fb-c13c17c7251c\"" -- set > >>>>>> Interface vnet0 external-ids:iface-status=active > >>>>>> > >>>>>> sb db after correcting the port: > >>>>>> > >>>>>> Chassis "6e4dd29f-7607-48d7-8e5a-eef4c6aeefb5" > >>>>>> hostname: "h2.limetransit.com" > >>>>>> Encap geneve > >>>>>> ip: "148.251.126.50" > >>>>>> options: {csum="true"} > >>>>>> Port_Binding "4981ee5f-6e15-4bd5-a1cf-7ead9bdd5873" > >>>>>> Port_Binding "ovirtbridge-port1" > >>>>>> Chassis "4f10fb04-8fb2-48d7-8a3f-ea6444c02cf9" > >>>>>> hostname: "h1.limetransit.com" > >>>>>> Encap geneve > >>>>>> ip: "144.76.84.73" > >>>>>> options: {csum="true"} > >>>>>> Port_Binding "ovirtbridge-port2" > >>>>>> Port_Binding "92f6d3c8-68b3-4986-9c09-60bee04644b5" > >>>>>> > >>>>>> I don't know from where the faulty id comes from, it's not in any > >>>>>> logs. > >>>>>> In the domain xml as printed in vdsm.log the id is correct: > >>>>>> > >>>>>> <interface type="bridge"> > >>>>>> <mac address="00:1a:4a:16:01:52" /> > >>>>>> <model type="virtio" /> > >>>>>> <source bridge="br-int" /> > >>>>>> <virtualport type="openvswitch" /> > >>>>>> <link state="up" /> > >>>>>> <boot order="2" /> > >>>>>> <bandwidth /> > >>>>>> <virtualport type="openvswitch"> > >>>>>> <parameters > >>>>>> interfaceid="92f6d3c8-68b3-4986-9c09-60bee04644b5" /> > >>>>>> </virtualport> > >>>>>> </interface> > >>>>>> > >>>>>> Where is the ovs-vsctl command line built for this call? > >>>>>> > >>>>>> /Sverker > >>>>>> > >>>>>> > >>>>>> Den 2017-01-02 kl. 13:40, skrev Sverker Abrahamsson: > >>>>>>> Got it to work now by following the env8 example in OVN tutorial, > >>>>>>> where a port is added with type l2gateway. Not sure how that is > >>>>>>> different from the localnet variant, but didn't suceed in > >>>>>>> getting that > >>>>>>> one working. Now I'm able to ping and telnet over the tunnel, > >>>>>>> but not > >>>>>>> ssh even when the port is answering on telnet. Neither does nfs > >>>>>>> traffic work even though mount did. Suspecting MTU issue. I did > >>>>>>> notice > >>>>>>> that ovn-controller starts too early, before network interfaces are > >>>>>>> established and hence can't reach the db. As these is a purely > >>>>>>> OVS/OVN > >>>>>>> issue I'll ask about it on their mailing list. > >>>>>>> > >>>>>>> Getting back to the original issue with Ovirt, I've now added the > >>>>>>> second host h1 to ovirt-engine. Had to do the same as with h2 to > >>>>>>> create a dummy ovirtmgmt network but configured access via the > >>>>>>> public > >>>>>>> IP. My firewall settings was replaced with iptables config and > >>>>>>> vdsm.conf was overwritten when engine was set up, so those had > >>>>>>> to be > >>>>>>> manually restored. It would be preferable if it would be > >>>>>>> possible to > >>>>>>> configure ovirt-engine that it does not "own" the host and instead > >>>>>>> comply with the settings it has instead of enforcing it's own > >>>>>>> view.. > >>>>>>> > >>>>>>> Apart from that it seems the second host works, although I need to > >>>>>>> resolve the traffic issue over the OVS tunnel. > >>>>>>> /Sverker > >>>>>>> > >>>>>>> Den 2017-01-02 kl. 01:13, skrev Sverker Abrahamsson: > >>>>>>>> 1. That is not possible as ovirt (or vdsm) will rewrite the > >>>>>>>> network > >>>>>>>> configuration to a non-working state. That is why I've set that > >>>>>>>> if as > >>>>>>>> hidden to vdsm and is why I'm keen on getting OVS/OVN to work > >>>>>>>> > >>>>>>>> 2. I've been reading the doc for OVN and starting to connect the > >>>>>>>> dots, which is not trivial as it is complex. Some insights > >>>>>>>> reached: > >>>>>>>> > >>>>>>>> First step is the OVN database, installed by > >>>>>>>> openvswitch-ovn-central, > >>>>>>>> which I currently have running on h2 host. The 'ovn-nbctl' and > >>>>>>>> 'ovn-sbctl' commands are only possible to execute on a database > >>>>>>>> node. > >>>>>>>> Two ip's are given to 'vdsm-tool ovn-config <ip to database> > >>>>>>>> <tunnel > >>>>>>>> ip>' as arguments, where <ip to database> is how this OVN node > >>>>>>>> reaches the database and <tunnel ip> is the ip to which other OVN > >>>>>>>> nodes sets up a tunnel to this node. I.e. it is not for creating a > >>>>>>>> tunnel to the database which I thought first from the > >>>>>>>> description in > >>>>>>>> blog post. > >>>>>>>> > >>>>>>>> The tunnel between OVN nodes is of type geneve which is a UDP > >>>>>>>> based > >>>>>>>> protocol but I have not been able to find anywhere which port > >>>>>>>> is used > >>>>>>>> so that I can open it in firewalld. I have added OVN on another > >>>>>>>> host, > >>>>>>>> called h1, and connected it to the db. I see there is traffic > >>>>>>>> to the > >>>>>>>> db port, but I don't see any geneve traffic between the nodes. > >>>>>>>> > >>>>>>>> Ovirt is now able to create it's vnet0 interface on the br-int ovs > >>>>>>>> bridge, but then I run into the next issue. How do I create a > >>>>>>>> connection from the logical switch to the physical host? I need > >>>>>>>> that > >>>>>>>> to a) get a connection out to the internet through a > >>>>>>>> masqueraded if > >>>>>>>> or ipv6 and b) be able to run a dhcp server to give ip's to the > >>>>>>>> VM's. > >>>>>>>> > >>>>>>>> /Sverker > >>>>>>>> > >>>>>>>> Den 2016-12-30 kl. 18:05, skrev Marcin Mirecki: > >>>>>>>>> 1. Why not use your physical nic for ovirtmgmt then? > >>>>>>>>> > >>>>>>>>> 2. "ovn-nbctl ls-add" does not add a bridge, but a logical > >>>>>>>>> switch. > >>>>>>>>> br-int is an internal OVN implementation detail, which > >>>>>>>>> the user > >>>>>>>>> should not care about. What you see in the ovirt UI are > >>>>>>>>> logical > >>>>>>>>> networks. They are implemented as OVN logical switches > >>>>>>>>> in case > >>>>>>>>> of the OVN provider. > >>>>>>>>> > >>>>>>>>> Please look at: > >>>>>>>>> http://www.ovirt.org/blog/2016/11/ovirt-provider-ovn/ > >>>>>>>>> You can get the latest rpms from here: > >>>>>>>>> http://resources.ovirt.org/repos/ovirt/experimental/master/ovirt-provider-ovn_fc24_46/rpm/fc24/noarch/ > >>>>>>>>> > >>>>>>>>> > >>>>>>>>> > >>>>>>>>> ----- Original Message ----- > >>>>>>>>>> From: "Sverker Abrahamsson"<sver...@abrahamsson.com> > >>>>>>>>>> To: "Marcin Mirecki"<mmire...@redhat.com> > >>>>>>>>>> Cc: "Ovirt Users"<users@ovirt.org> > >>>>>>>>>> Sent: Friday, December 30, 2016 4:25:58 PM > >>>>>>>>>> Subject: Re: [ovirt-users] Issue with OVN/OVS and mandatory > >>>>>>>>>> ovirtmgmt network > >>>>>>>>>> > >>>>>>>>>> 1. No, I did not want to put the ovirtmgmt bridge on my physical > >>>>>>>>>> nic as > >>>>>>>>>> it always messed up the network config making the host > >>>>>>>>>> unreachable. I > >>>>>>>>>> have put a ovs bridge on this nic which I will use to make > >>>>>>>>>> tunnels > >>>>>>>>>> when > >>>>>>>>>> I add other hosts. Maybe br-int will be used for that > >>>>>>>>>> instead, will > >>>>>>>>>> see > >>>>>>>>>> when I get that far. > >>>>>>>>>> > >>>>>>>>>> As it is now I have a dummy if for ovirtmgmt bridge but this > >>>>>>>>>> will > >>>>>>>>>> probably not work when I add other hosts as that bridge cannot > >>>>>>>>>> connect > >>>>>>>>>> to the other hosts. I'm considering keeping this just as a > >>>>>>>>>> dummy to > >>>>>>>>>> keep > >>>>>>>>>> ovirt engine satisfied while the actual communication will > >>>>>>>>>> happen > >>>>>>>>>> over > >>>>>>>>>> OVN/OVS bridges and tunnels. > >>>>>>>>>> > >>>>>>>>>> 2. On > >>>>>>>>>> https://www.ovirt.org//develop/release-management/features/ovirt-ovn-provider/ > >>>>>>>>>> > >>>>>>>>>> > >>>>>>>>>> there is instructions how to add an OVS bridge to OVN with > >>>>>>>>>> |ovn-nbctl > >>>>>>>>>> ls-add <network name>|. If you want to use br-int then it makes > >>>>>>>>>> sense to > >>>>>>>>>> make that bridge visible in ovirt webui under networks so > >>>>>>>>>> that it > >>>>>>>>>> can be > >>>>>>>>>> selected for VM's. > >>>>>>>>>> > >>>>>>>>>> It quite doesn't make sense to me that I can select other > >>>>>>>>>> network > >>>>>>>>>> for my > >>>>>>>>>> VM but then that setting is not used when setting up the > >>>>>>>>>> network. > >>>>>>>>>> > >>>>>>>>>> /Sverker > >>>>>>>>>> > >>>>>>>>>> Den 2016-12-30 kl. 15:34, skrev Marcin Mirecki: > >>>>>>>>>>> Hi, > >>>>>>>>>>> > >>>>>>>>>>> The OVN provider does not require you to add any bridges > >>>>>>>>>>> manually. > >>>>>>>>>>> As I understand we were dealing with two problems: > >>>>>>>>>>> 1. You only had one physical nic and wanted to put a bridge > >>>>>>>>>>> on it, > >>>>>>>>>>> attaching the management network to the bridge. This > >>>>>>>>>>> was the > >>>>>>>>>>> reason for > >>>>>>>>>>> creating the bridge (the recommended setup would be > >>>>>>>>>>> to used a > >>>>>>>>>>> separate > >>>>>>>>>>> physical nic for the management network). This bridge > >>>>>>>>>>> has > >>>>>>>>>>> nothing to > >>>>>>>>>>> do with the OVN bridge. > >>>>>>>>>>> 2. OVN - you want to use OVN on this system. For this you > >>>>>>>>>>> have to > >>>>>>>>>>> install > >>>>>>>>>>> OVN on your hosts. This should create the br-int bridge, > >>>>>>>>>>> which are > >>>>>>>>>>> then used by the OVN provider. This br-int bridge > >>>>>>>>>>> must be > >>>>>>>>>>> configured > >>>>>>>>>>> to connect to other hosts using the geneve tunnels. > >>>>>>>>>>> > >>>>>>>>>>> In both cases the systems will not be aware of any bridges you > >>>>>>>>>>> create. > >>>>>>>>>>> They need a nic (be it physical or virtual) to connect to other > >>>>>>>>>>> system. > >>>>>>>>>>> Usually this is the physical nic. In your case you decided > >>>>>>>>>>> to put > >>>>>>>>>>> a bridge > >>>>>>>>>>> on the physical nic, and give oVirt a virtual nic attached > >>>>>>>>>>> to this > >>>>>>>>>>> bridge. > >>>>>>>>>>> This works, but keep in mind that the bridge you have > >>>>>>>>>>> introduced > >>>>>>>>>>> is outside > >>>>>>>>>>> of oVirt's (and OVN) control (and as such is not supported). > >>>>>>>>>>> > >>>>>>>>>>>> What is the purpose of > >>>>>>>>>>>> adding my bridges to Ovirt through the external provider and > >>>>>>>>>>>> configure > >>>>>>>>>>>> them on my VM > >>>>>>>>>>> I am not quite sure I understand. > >>>>>>>>>>> The external provider (OVN provider to be specific), does > >>>>>>>>>>> not add > >>>>>>>>>>> any > >>>>>>>>>>> bridges > >>>>>>>>>>> to the system. It is using the br-int bridge created by OVN. > >>>>>>>>>>> The > >>>>>>>>>>> networks > >>>>>>>>>>> created by the OVN provider are purely logical entities, > >>>>>>>>>>> implemented using > >>>>>>>>>>> the OVN br-int bridge. > >>>>>>>>>>> > >>>>>>>>>>> Marcin > >>>>>>>>>>> > >>>>>>>>>>> > >>>>>>>>>>> ----- Original Message ----- > >>>>>>>>>>>> From: "Sverker Abrahamsson"<sver...@abrahamsson.com> > >>>>>>>>>>>> To: "Marcin Mirecki"<mmire...@redhat.com> > >>>>>>>>>>>> Cc: "Ovirt Users"<users@ovirt.org> > >>>>>>>>>>>> Sent: Friday, December 30, 2016 12:15:43 PM > >>>>>>>>>>>> Subject: Re: [ovirt-users] Issue with OVN/OVS and mandatory > >>>>>>>>>>>> ovirtmgmt > >>>>>>>>>>>> network > >>>>>>>>>>>> > >>>>>>>>>>>> Hi > >>>>>>>>>>>> That is the logic I quite don't understand. What is the > >>>>>>>>>>>> purpose of > >>>>>>>>>>>> adding my bridges to Ovirt through the external provider and > >>>>>>>>>>>> configure > >>>>>>>>>>>> them on my VM if you are disregarding that and using br-int > >>>>>>>>>>>> anyway? > >>>>>>>>>>>> > >>>>>>>>>>>> /Sverker > >>>>>>>>>>>> > >>>>>>>>>>>> Den 2016-12-30 kl. 10:53, skrev Marcin Mirecki: > >>>>>>>>>>>>> Sverker, > >>>>>>>>>>>>> > >>>>>>>>>>>>> br-int is the integration bridge created by default in > >>>>>>>>>>>>> OVN. This > >>>>>>>>>>>>> is the > >>>>>>>>>>>>> bridge we use for the OVN provider. As OVN is required to be > >>>>>>>>>>>>> installed, > >>>>>>>>>>>>> we assume that this bridge is present. > >>>>>>>>>>>>> Using any other ovs bridge is not supported, and will require > >>>>>>>>>>>>> custom code > >>>>>>>>>>>>> changes (such as the ones you created). > >>>>>>>>>>>>> > >>>>>>>>>>>>> The proper setup in your case would probably be to create > >>>>>>>>>>>>> br-int > >>>>>>>>>>>>> and > >>>>>>>>>>>>> connect > >>>>>>>>>>>>> this to your ovirtbridge, although I don't know the > >>>>>>>>>>>>> details of > >>>>>>>>>>>>> your env, > >>>>>>>>>>>>> so > >>>>>>>>>>>>> this is just my best guess. > >>>>>>>>>>>>> > >>>>>>>>>>>>> Marcin > >>>>>>>>>>>>> > >>>>>>>>>>>>> > >>>>>>>>>>>>> ----- Original Message ----- > >>>>>>>>>>>>>> From: "Sverker Abrahamsson"<sver...@abrahamsson.com> > >>>>>>>>>>>>>> To: "Marcin Mirecki"<mmire...@redhat.com> > >>>>>>>>>>>>>> Cc: "Ovirt Users"<users@ovirt.org>, "Numan Siddique" > >>>>>>>>>>>>>> <nusid...@redhat.com> > >>>>>>>>>>>>>> Sent: Friday, December 30, 2016 1:14:50 AM > >>>>>>>>>>>>>> Subject: Re: [ovirt-users] Issue with OVN/OVS and mandatory > >>>>>>>>>>>>>> ovirtmgmt > >>>>>>>>>>>>>> network > >>>>>>>>>>>>>> > >>>>>>>>>>>>>> Even better, if the value is not hardcoded then the > >>>>>>>>>>>>>> configured > >>>>>>>>>>>>>> value is > >>>>>>>>>>>>>> used. Might be that I'm missunderstanding something but > >>>>>>>>>>>>>> this is > >>>>>>>>>>>>>> the > >>>>>>>>>>>>>> behaviour I expected instead of that it is using br-int. > >>>>>>>>>>>>>> > >>>>>>>>>>>>>> Attached is a patch which properly sets up the xml, in case > >>>>>>>>>>>>>> there is > >>>>>>>>>>>>>> already a virtual port there + testcode of some variants > >>>>>>>>>>>>>> > >>>>>>>>>>>>>> /Sverker > >>>>>>>>>>>>>> > >>>>>>>>>>>>>> Den 2016-12-29 kl. 22:55, skrev Sverker Abrahamsson: > >>>>>>>>>>>>>>> When I change > >>>>>>>>>>>>>>> /usr/libexec/vdsm/hooks/before_device_create/ovirt_provider_ovn_hook > >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> to instead of hardcoded to br-int use BRIDGE_NAME = > >>>>>>>>>>>>>>> 'ovirtbridge' then > >>>>>>>>>>>>>>> I get the expected behaviour and I get a working network > >>>>>>>>>>>>>>> connectivity > >>>>>>>>>>>>>>> in my VM with IP provided by dhcp. > >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> /Sverker > >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> Den 2016-12-29 kl. 22:07, skrev Sverker Abrahamsson: > >>>>>>>>>>>>>>>> By default the vNic profile of my OVN bridge > >>>>>>>>>>>>>>>> ovirtbridge gets a > >>>>>>>>>>>>>>>> Network filter named vdsm-no-mac-spoofing. If I instead > >>>>>>>>>>>>>>>> set > >>>>>>>>>>>>>>>> No filter > >>>>>>>>>>>>>>>> then I don't get those ebtables / iptables messages. It > >>>>>>>>>>>>>>>> seems > >>>>>>>>>>>>>>>> that > >>>>>>>>>>>>>>>> there is some issue between ovirt/vdsm and firewalld, > >>>>>>>>>>>>>>>> which > >>>>>>>>>>>>>>>> we can > >>>>>>>>>>>>>>>> put to the side for now. > >>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>> It is not clear for me why the port is added on br-int > >>>>>>>>>>>>>>>> instead of the > >>>>>>>>>>>>>>>> bridge I've assigned to the VM, which is ovirtbridge?? > >>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>> /Sverker > >>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>> Den 2016-12-29 kl. 14:20, skrev Sverker Abrahamsson: > >>>>>>>>>>>>>>>>> The specific command most likely fails because there > >>>>>>>>>>>>>>>>> is no > >>>>>>>>>>>>>>>>> chain > >>>>>>>>>>>>>>>>> named libvirt-J-vnet0, but when should that have been > >>>>>>>>>>>>>>>>> created? > >>>>>>>>>>>>>>>>> /Sverker > >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> -------- Vidarebefordrat meddelande -------- > >>>>>>>>>>>>>>>>> Ämne: Re: [ovirt-users] Issue with OVN/OVS and > >>>>>>>>>>>>>>>>> mandatory > >>>>>>>>>>>>>>>>> ovirtmgmt > >>>>>>>>>>>>>>>>> network > >>>>>>>>>>>>>>>>> Datum: Thu, 29 Dec 2016 08:06:29 -0500 (EST) > >>>>>>>>>>>>>>>>> Från: Marcin Mirecki<mmire...@redhat.com> > >>>>>>>>>>>>>>>>> Till: Sverker Abrahamsson<sver...@abrahamsson.com> > >>>>>>>>>>>>>>>>> Kopia: Ovirt Users<users@ovirt.org>, Lance Richardson > >>>>>>>>>>>>>>>>> <lrich...@redhat.com>, Numan > >>>>>>>>>>>>>>>>> Siddique<nusid...@redhat.com> > >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> Let me add the OVN team. > >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> Lance, Numan, > >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> Can you please look at this? > >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> Trying to plug a vNIC results in: > >>>>>>>>>>>>>>>>>>>>>>>> Dec 28 23:31:35 h2 ovs-vsctl: > >>>>>>>>>>>>>>>>>>>>>>>> ovs|00001|vsctl|INFO|Called as > >>>>>>>>>>>>>>>>>>>>>>>> ovs-vsctl > >>>>>>>>>>>>>>>>>>>>>>>> --timeout=5 -- --if-exists del-port vnet0 -- > >>>>>>>>>>>>>>>>>>>>>>>> add-port > >>>>>>>>>>>>>>>>>>>>>>>> br-int > >>>>>>>>>>>>>>>>>>>>>>>> vnet0 -- > >>>>>>>>>>>>>>>>>>>>>>>> set Interface vnet0 > >>>>>>>>>>>>>>>>>>>>>>>> "external-ids:attached-mac=\"00:1a:4a:16:01:51\"" > >>>>>>>>>>>>>>>>>>>>>>>> -- set Interface vnet0 > >>>>>>>>>>>>>>>>>>>>>>>> "external-ids:iface-id=\"e8853aac-8a75-41b0-8010-e630017dcdd8\"" > >>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>> -- > >>>>>>>>>>>>>>>>>>>>>>>> set Interface vnet0 > >>>>>>>>>>>>>>>>>>>>>>>> "external-ids:vm-id=\"b9440d60-ef5a-4e2b-83cf-081df7c09e6f\"" > >>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>> -- > >>>>>>>>>>>>>>>>>>>>>>>> set > >>>>>>>>>>>>>>>>>>>>>>>> Interface vnet0 external-ids:iface-status=active > >>>>>>>>>>>>>>>>>>>>>>>> Dec 28 23:31:35 h2 kernel: device vnet0 entered > >>>>>>>>>>>>>>>>>>>>>>>> promiscuous > >>>>>>>>>>>>>>>>>>>>>>>> mode > >>>>>>>>>>>>>>>>>>>>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: > >>>>>>>>>>>>>>>>>>>>>>>> COMMAND_FAILED: > >>>>>>>>>>>>>>>>>>>>>>>> '/usr/sbin/ebtables --concurrent -t nat -D > >>>>>>>>>>>>>>>>>>>>>>>> PREROUTING > >>>>>>>>>>>>>>>>>>>>>>>> -i vnet0 > >>>>>>>>>>>>>>>>>>>>>>>> -j > >>>>>>>>>>>>>>>>>>>>>>>> libvirt-J-vnet0' failed: > >>>>>>>>>>>>>>>>>>>>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: > >>>>>>>>>>>>>>>>>>>>>>>> COMMAND_FAILED: > >>>>>>>>>>>>>>>>> More details below > >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> ----- Original Message ----- > >>>>>>>>>>>>>>>>>> From: "Sverker Abrahamsson"<sver...@abrahamsson.com> > >>>>>>>>>>>>>>>>>> To: "Marcin Mirecki"<mmire...@redhat.com> > >>>>>>>>>>>>>>>>>> Cc: "Ovirt Users"<users@ovirt.org> > >>>>>>>>>>>>>>>>>> Sent: Thursday, December 29, 2016 1:42:11 PM > >>>>>>>>>>>>>>>>>> Subject: Re: [ovirt-users] Issue with OVN/OVS and > >>>>>>>>>>>>>>>>>> mandatory > >>>>>>>>>>>>>>>>>> ovirtmgmt > >>>>>>>>>>>>>>>>>> network > >>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>> Hi > >>>>>>>>>>>>>>>>>> Same problem still.. > >>>>>>>>>>>>>>>>>> /Sverker > >>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>> Den 2016-12-29 kl. 13:34, skrev Marcin Mirecki: > >>>>>>>>>>>>>>>>>>> Hi, > >>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>> The tunnels are created to connect multiple OVN > >>>>>>>>>>>>>>>>>>> controllers. > >>>>>>>>>>>>>>>>>>> If there is only one, there is no need for the > >>>>>>>>>>>>>>>>>>> tunnels, so > >>>>>>>>>>>>>>>>>>> none > >>>>>>>>>>>>>>>>>>> will be created, this is the correct behavior. > >>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>> Does the problem still occur after setting > >>>>>>>>>>>>>>>>>>> configuring the > >>>>>>>>>>>>>>>>>>> OVN-controller? > >>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>> Marcin > >>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>> ----- Original Message ----- > >>>>>>>>>>>>>>>>>>>> From: "Sverker Abrahamsson"<sver...@abrahamsson.com> > >>>>>>>>>>>>>>>>>>>> To: "Marcin Mirecki"<mmire...@redhat.com> > >>>>>>>>>>>>>>>>>>>> Cc: "Ovirt Users"<users@ovirt.org> > >>>>>>>>>>>>>>>>>>>> Sent: Thursday, December 29, 2016 11:44:32 AM > >>>>>>>>>>>>>>>>>>>> Subject: Re: [ovirt-users] Issue with OVN/OVS and > >>>>>>>>>>>>>>>>>>>> mandatory > >>>>>>>>>>>>>>>>>>>> ovirtmgmt > >>>>>>>>>>>>>>>>>>>> network > >>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>> Hi > >>>>>>>>>>>>>>>>>>>> The rpm packages you listed in the other mail are > >>>>>>>>>>>>>>>>>>>> installed but I > >>>>>>>>>>>>>>>>>>>> had > >>>>>>>>>>>>>>>>>>>> not run vdsm-tool ovn-config to create tunnel as > >>>>>>>>>>>>>>>>>>>> the OVN > >>>>>>>>>>>>>>>>>>>> controller > >>>>>>>>>>>>>>>>>>>> is > >>>>>>>>>>>>>>>>>>>> on the same host. > >>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>> [root@h2 ~]# rpm -q openvswitch-ovn-common > >>>>>>>>>>>>>>>>>>>> openvswitch-ovn-common-2.6.90-1.el7.centos.x86_64 > >>>>>>>>>>>>>>>>>>>> [root@h2 ~]# rpm -q openvswitch-ovn-host > >>>>>>>>>>>>>>>>>>>> openvswitch-ovn-host-2.6.90-1.el7.centos.x86_64 > >>>>>>>>>>>>>>>>>>>> [root@h2 ~]# rpm -q python-openvswitch > >>>>>>>>>>>>>>>>>>>> python-openvswitch-2.6.90-1.el7.centos.noarch > >>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>> After removing my manually created br-int and run > >>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>> vdsm-tool ovn-config 127.0.0.1 172.27.1.1 > >>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>> then I have the br-int but 'ip link show' does not > >>>>>>>>>>>>>>>>>>>> show > >>>>>>>>>>>>>>>>>>>> any link > >>>>>>>>>>>>>>>>>>>> 'genev_sys_' nor does 'ovs-vsctl show' any port for > >>>>>>>>>>>>>>>>>>>> ovn. > >>>>>>>>>>>>>>>>>>>> I assume > >>>>>>>>>>>>>>>>>>>> these > >>>>>>>>>>>>>>>>>>>> are when there is an actual tunnel? > >>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>> [root@h2 ~]# ovs-vsctl show > >>>>>>>>>>>>>>>>>>>> ebb6aede-cbbc-4f4f-a88a-a9cd72b2bd23 > >>>>>>>>>>>>>>>>>>>> Bridge br-int > >>>>>>>>>>>>>>>>>>>> fail_mode: secure > >>>>>>>>>>>>>>>>>>>> Port br-int > >>>>>>>>>>>>>>>>>>>> Interface br-int > >>>>>>>>>>>>>>>>>>>> type: internal > >>>>>>>>>>>>>>>>>>>> Bridge ovirtbridge > >>>>>>>>>>>>>>>>>>>> Port ovirtbridge > >>>>>>>>>>>>>>>>>>>> Interface ovirtbridge > >>>>>>>>>>>>>>>>>>>> type: internal > >>>>>>>>>>>>>>>>>>>> Bridge "ovsbridge0" > >>>>>>>>>>>>>>>>>>>> Port "ovsbridge0" > >>>>>>>>>>>>>>>>>>>> Interface "ovsbridge0" > >>>>>>>>>>>>>>>>>>>> type: internal > >>>>>>>>>>>>>>>>>>>> Port "eth0" > >>>>>>>>>>>>>>>>>>>> Interface "eth0" > >>>>>>>>>>>>>>>>>>>> ovs_version: "2.6.90" > >>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>> [root@h2 ~]# ip link show > >>>>>>>>>>>>>>>>>>>> 1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc > >>>>>>>>>>>>>>>>>>>> noqueue state > >>>>>>>>>>>>>>>>>>>> UNKNOWN > >>>>>>>>>>>>>>>>>>>> mode > >>>>>>>>>>>>>>>>>>>> DEFAULT qlen 1 > >>>>>>>>>>>>>>>>>>>> link/loopback 00:00:00:00:00:00 brd > >>>>>>>>>>>>>>>>>>>> 00:00:00:00:00:00 > >>>>>>>>>>>>>>>>>>>> 2: eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 > >>>>>>>>>>>>>>>>>>>> qdisc > >>>>>>>>>>>>>>>>>>>> pfifo_fast > >>>>>>>>>>>>>>>>>>>> master ovs-system state UP mode DEFAULT qlen 1000 > >>>>>>>>>>>>>>>>>>>> link/ether 44:8a:5b:84:7d:b3 brd > >>>>>>>>>>>>>>>>>>>> ff:ff:ff:ff:ff:ff > >>>>>>>>>>>>>>>>>>>> 3: ovs-system: <BROADCAST,MULTICAST> mtu 1500 qdisc > >>>>>>>>>>>>>>>>>>>> noop > >>>>>>>>>>>>>>>>>>>> state > >>>>>>>>>>>>>>>>>>>> DOWN > >>>>>>>>>>>>>>>>>>>> mode > >>>>>>>>>>>>>>>>>>>> DEFAULT qlen 1000 > >>>>>>>>>>>>>>>>>>>> link/ether 5a:14:cf:28:47:e2 brd > >>>>>>>>>>>>>>>>>>>> ff:ff:ff:ff:ff:ff > >>>>>>>>>>>>>>>>>>>> 4: ovsbridge0: <BROADCAST,MULTICAST,UP,LOWER_UP> > >>>>>>>>>>>>>>>>>>>> mtu 1500 > >>>>>>>>>>>>>>>>>>>> qdisc > >>>>>>>>>>>>>>>>>>>> noqueue > >>>>>>>>>>>>>>>>>>>> state UNKNOWN mode DEFAULT qlen 1000 > >>>>>>>>>>>>>>>>>>>> link/ether 44:8a:5b:84:7d:b3 brd > >>>>>>>>>>>>>>>>>>>> ff:ff:ff:ff:ff:ff > >>>>>>>>>>>>>>>>>>>> 5: br-int: <BROADCAST,MULTICAST> mtu 1500 qdisc noop > >>>>>>>>>>>>>>>>>>>> state DOWN > >>>>>>>>>>>>>>>>>>>> mode > >>>>>>>>>>>>>>>>>>>> DEFAULT qlen 1000 > >>>>>>>>>>>>>>>>>>>> link/ether 9e:b0:3a:9d:f2:4b brd > >>>>>>>>>>>>>>>>>>>> ff:ff:ff:ff:ff:ff > >>>>>>>>>>>>>>>>>>>> 6: ovirtbridge: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu > >>>>>>>>>>>>>>>>>>>> 1500 qdisc > >>>>>>>>>>>>>>>>>>>> noqueue > >>>>>>>>>>>>>>>>>>>> state UNKNOWN mode DEFAULT qlen 1000 > >>>>>>>>>>>>>>>>>>>> link/ether a6:f6:e5:a4:5b:45 brd > >>>>>>>>>>>>>>>>>>>> ff:ff:ff:ff:ff:ff > >>>>>>>>>>>>>>>>>>>> 7: dummy0: <BROADCAST,NOARP,UP,LOWER_UP> mtu 1500 > >>>>>>>>>>>>>>>>>>>> qdisc > >>>>>>>>>>>>>>>>>>>> noqueue > >>>>>>>>>>>>>>>>>>>> master > >>>>>>>>>>>>>>>>>>>> ovirtmgmt state UNKNOWN mode DEFAULT qlen 1000 > >>>>>>>>>>>>>>>>>>>> link/ether 66:e0:1c:c3:a9:d8 brd > >>>>>>>>>>>>>>>>>>>> ff:ff:ff:ff:ff:ff > >>>>>>>>>>>>>>>>>>>> 8: ovirtmgmt: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu > >>>>>>>>>>>>>>>>>>>> 1500 > >>>>>>>>>>>>>>>>>>>> qdisc > >>>>>>>>>>>>>>>>>>>> noqueue > >>>>>>>>>>>>>>>>>>>> state UP mode DEFAULT qlen 1000 > >>>>>>>>>>>>>>>>>>>> link/ether 66:e0:1c:c3:a9:d8 brd > >>>>>>>>>>>>>>>>>>>> ff:ff:ff:ff:ff:ff > >>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>> Firewall settings: > >>>>>>>>>>>>>>>>>>>> [root@h2 ~]# firewall-cmd --list-all-zones > >>>>>>>>>>>>>>>>>>>> work > >>>>>>>>>>>>>>>>>>>> target: default > >>>>>>>>>>>>>>>>>>>> icmp-block-inversion: no > >>>>>>>>>>>>>>>>>>>> interfaces: > >>>>>>>>>>>>>>>>>>>> sources: > >>>>>>>>>>>>>>>>>>>> services: dhcpv6-client ssh > >>>>>>>>>>>>>>>>>>>> ports: > >>>>>>>>>>>>>>>>>>>> protocols: > >>>>>>>>>>>>>>>>>>>> masquerade: no > >>>>>>>>>>>>>>>>>>>> forward-ports: > >>>>>>>>>>>>>>>>>>>> sourceports: > >>>>>>>>>>>>>>>>>>>> icmp-blocks: > >>>>>>>>>>>>>>>>>>>> rich rules: > >>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>> drop > >>>>>>>>>>>>>>>>>>>> target: DROP > >>>>>>>>>>>>>>>>>>>> icmp-block-inversion: no > >>>>>>>>>>>>>>>>>>>> interfaces: > >>>>>>>>>>>>>>>>>>>> sources: > >>>>>>>>>>>>>>>>>>>> services: > >>>>>>>>>>>>>>>>>>>> ports: > >>>>>>>>>>>>>>>>>>>> protocols: > >>>>>>>>>>>>>>>>>>>> masquerade: no > >>>>>>>>>>>>>>>>>>>> forward-ports: > >>>>>>>>>>>>>>>>>>>> sourceports: > >>>>>>>>>>>>>>>>>>>> icmp-blocks: > >>>>>>>>>>>>>>>>>>>> rich rules: > >>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>> internal > >>>>>>>>>>>>>>>>>>>> target: default > >>>>>>>>>>>>>>>>>>>> icmp-block-inversion: no > >>>>>>>>>>>>>>>>>>>> interfaces: > >>>>>>>>>>>>>>>>>>>> sources: > >>>>>>>>>>>>>>>>>>>> services: dhcpv6-client mdns samba-client ssh > >>>>>>>>>>>>>>>>>>>> ports: > >>>>>>>>>>>>>>>>>>>> protocols: > >>>>>>>>>>>>>>>>>>>> masquerade: no > >>>>>>>>>>>>>>>>>>>> forward-ports: > >>>>>>>>>>>>>>>>>>>> sourceports: > >>>>>>>>>>>>>>>>>>>> icmp-blocks: > >>>>>>>>>>>>>>>>>>>> rich rules: > >>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>> external > >>>>>>>>>>>>>>>>>>>> target: default > >>>>>>>>>>>>>>>>>>>> icmp-block-inversion: no > >>>>>>>>>>>>>>>>>>>> interfaces: > >>>>>>>>>>>>>>>>>>>> sources: > >>>>>>>>>>>>>>>>>>>> services: ssh > >>>>>>>>>>>>>>>>>>>> ports: > >>>>>>>>>>>>>>>>>>>> protocols: > >>>>>>>>>>>>>>>>>>>> masquerade: yes > >>>>>>>>>>>>>>>>>>>> forward-ports: > >>>>>>>>>>>>>>>>>>>> sourceports: > >>>>>>>>>>>>>>>>>>>> icmp-blocks: > >>>>>>>>>>>>>>>>>>>> rich rules: > >>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>> trusted > >>>>>>>>>>>>>>>>>>>> target: ACCEPT > >>>>>>>>>>>>>>>>>>>> icmp-block-inversion: no > >>>>>>>>>>>>>>>>>>>> interfaces: > >>>>>>>>>>>>>>>>>>>> sources: > >>>>>>>>>>>>>>>>>>>> services: > >>>>>>>>>>>>>>>>>>>> ports: > >>>>>>>>>>>>>>>>>>>> protocols: > >>>>>>>>>>>>>>>>>>>> masquerade: no > >>>>>>>>>>>>>>>>>>>> forward-ports: > >>>>>>>>>>>>>>>>>>>> sourceports: > >>>>>>>>>>>>>>>>>>>> icmp-blocks: > >>>>>>>>>>>>>>>>>>>> rich rules: > >>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>> home > >>>>>>>>>>>>>>>>>>>> target: default > >>>>>>>>>>>>>>>>>>>> icmp-block-inversion: no > >>>>>>>>>>>>>>>>>>>> interfaces: > >>>>>>>>>>>>>>>>>>>> sources: > >>>>>>>>>>>>>>>>>>>> services: dhcpv6-client mdns samba-client ssh > >>>>>>>>>>>>>>>>>>>> ports: > >>>>>>>>>>>>>>>>>>>> protocols: > >>>>>>>>>>>>>>>>>>>> masquerade: no > >>>>>>>>>>>>>>>>>>>> forward-ports: > >>>>>>>>>>>>>>>>>>>> sourceports: > >>>>>>>>>>>>>>>>>>>> icmp-blocks: > >>>>>>>>>>>>>>>>>>>> rich rules: > >>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>> dmz > >>>>>>>>>>>>>>>>>>>> target: default > >>>>>>>>>>>>>>>>>>>> icmp-block-inversion: no > >>>>>>>>>>>>>>>>>>>> interfaces: > >>>>>>>>>>>>>>>>>>>> sources: > >>>>>>>>>>>>>>>>>>>> services: ssh > >>>>>>>>>>>>>>>>>>>> ports: > >>>>>>>>>>>>>>>>>>>> protocols: > >>>>>>>>>>>>>>>>>>>> masquerade: no > >>>>>>>>>>>>>>>>>>>> forward-ports: > >>>>>>>>>>>>>>>>>>>> sourceports: > >>>>>>>>>>>>>>>>>>>> icmp-blocks: > >>>>>>>>>>>>>>>>>>>> rich rules: > >>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>> public (active) > >>>>>>>>>>>>>>>>>>>> target: default > >>>>>>>>>>>>>>>>>>>> icmp-block-inversion: no > >>>>>>>>>>>>>>>>>>>> interfaces: eth0 ovsbridge0 > >>>>>>>>>>>>>>>>>>>> sources: > >>>>>>>>>>>>>>>>>>>> services: dhcpv6-client ssh > >>>>>>>>>>>>>>>>>>>> ports: > >>>>>>>>>>>>>>>>>>>> protocols: > >>>>>>>>>>>>>>>>>>>> masquerade: no > >>>>>>>>>>>>>>>>>>>> forward-ports: > >>>>>>>>>>>>>>>>>>>> sourceports: > >>>>>>>>>>>>>>>>>>>> icmp-blocks: > >>>>>>>>>>>>>>>>>>>> rich rules: > >>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>> block > >>>>>>>>>>>>>>>>>>>> target: %%REJECT%% > >>>>>>>>>>>>>>>>>>>> icmp-block-inversion: no > >>>>>>>>>>>>>>>>>>>> interfaces: > >>>>>>>>>>>>>>>>>>>> sources: > >>>>>>>>>>>>>>>>>>>> services: > >>>>>>>>>>>>>>>>>>>> ports: > >>>>>>>>>>>>>>>>>>>> protocols: > >>>>>>>>>>>>>>>>>>>> masquerade: no > >>>>>>>>>>>>>>>>>>>> forward-ports: > >>>>>>>>>>>>>>>>>>>> sourceports: > >>>>>>>>>>>>>>>>>>>> icmp-blocks: > >>>>>>>>>>>>>>>>>>>> rich rules: > >>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>> ovirt (active) > >>>>>>>>>>>>>>>>>>>> target: default > >>>>>>>>>>>>>>>>>>>> icmp-block-inversion: no > >>>>>>>>>>>>>>>>>>>> interfaces: ovirtbridge ovirtmgmt > >>>>>>>>>>>>>>>>>>>> sources: > >>>>>>>>>>>>>>>>>>>> services: dhcp ovirt-fence-kdump-listener > >>>>>>>>>>>>>>>>>>>> ovirt-http > >>>>>>>>>>>>>>>>>>>> ovirt-https > >>>>>>>>>>>>>>>>>>>> ovirt-imageio-proxy ovirt-postgres ovirt-provider-ovn > >>>>>>>>>>>>>>>>>>>> ovirt-vmconsole-proxy ovirt-websocket-proxy ssh vdsm > >>>>>>>>>>>>>>>>>>>> ports: > >>>>>>>>>>>>>>>>>>>> protocols: > >>>>>>>>>>>>>>>>>>>> masquerade: yes > >>>>>>>>>>>>>>>>>>>> forward-ports: > >>>>>>>>>>>>>>>>>>>> sourceports: > >>>>>>>>>>>>>>>>>>>> icmp-blocks: > >>>>>>>>>>>>>>>>>>>> rich rules: > >>>>>>>>>>>>>>>>>>>> rule family="ipv4" port port="6641" > >>>>>>>>>>>>>>>>>>>> protocol="tcp" > >>>>>>>>>>>>>>>>>>>> accept > >>>>>>>>>>>>>>>>>>>> rule family="ipv4" port port="6642" > >>>>>>>>>>>>>>>>>>>> protocol="tcp" > >>>>>>>>>>>>>>>>>>>> accept > >>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>> The db dump is attached > >>>>>>>>>>>>>>>>>>>> /Sverker > >>>>>>>>>>>>>>>>>>>> Den 2016-12-29 kl. 09:50, skrev Marcin Mirecki: > >>>>>>>>>>>>>>>>>>>>> Hi, > >>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>> Can you please do: "sudo ovsdb-client dump" > >>>>>>>>>>>>>>>>>>>>> on the host and send me the output? > >>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>> Have you configured the ovn controller to connect > >>>>>>>>>>>>>>>>>>>>> to the > >>>>>>>>>>>>>>>>>>>>> OVN north? You can do it using "vdsm-tool > >>>>>>>>>>>>>>>>>>>>> ovn-config" or > >>>>>>>>>>>>>>>>>>>>> using the OVN tools directly. > >>>>>>>>>>>>>>>>>>>>> Please check > >>>>>>>>>>>>>>>>>>>>> out:https://www.ovirt.org/blog/2016/11/ovirt-provider-ovn/ > >>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>> for details. > >>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>> Also please note that the OVN provider is completely > >>>>>>>>>>>>>>>>>>>>> different > >>>>>>>>>>>>>>>>>>>>> from the neutron-openvswitch plugin. Please don't mix > >>>>>>>>>>>>>>>>>>>>> the two. > >>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>> Marcin > >>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>> ----- Original Message ----- > >>>>>>>>>>>>>>>>>>>>>> From: "Marcin Mirecki"<mmire...@redhat.com> > >>>>>>>>>>>>>>>>>>>>>> To: "Sverker Abrahamsson"<sver...@abrahamsson.com> > >>>>>>>>>>>>>>>>>>>>>> Cc: "Ovirt Users"<users@ovirt.org> > >>>>>>>>>>>>>>>>>>>>>> Sent: Thursday, December 29, 2016 9:27:19 AM > >>>>>>>>>>>>>>>>>>>>>> Subject: Re: [ovirt-users] Issue with OVN/OVS and > >>>>>>>>>>>>>>>>>>>>>> mandatory > >>>>>>>>>>>>>>>>>>>>>> ovirtmgmt > >>>>>>>>>>>>>>>>>>>>>> network > >>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>> Hi, > >>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>> br-int is the OVN integration bridge, it should > >>>>>>>>>>>>>>>>>>>>>> have been > >>>>>>>>>>>>>>>>>>>>>> created > >>>>>>>>>>>>>>>>>>>>>> when installing OVN. I assume you have the following > >>>>>>>>>>>>>>>>>>>>>> packages > >>>>>>>>>>>>>>>>>>>>>> installed > >>>>>>>>>>>>>>>>>>>>>> on the host: > >>>>>>>>>>>>>>>>>>>>>> openvswitch-ovn-common > >>>>>>>>>>>>>>>>>>>>>> openvswitch-ovn-host > >>>>>>>>>>>>>>>>>>>>>> python-openvswitch > >>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>> Please give me some time to look at the connectivity > >>>>>>>>>>>>>>>>>>>>>> problem. > >>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>> Marcin > >>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>> ----- Original Message ----- > >>>>>>>>>>>>>>>>>>>>>>> From: "Sverker > >>>>>>>>>>>>>>>>>>>>>>> Abrahamsson"<sver...@abrahamsson.com> > >>>>>>>>>>>>>>>>>>>>>>> To: "Marcin Mirecki"<mmire...@redhat.com> > >>>>>>>>>>>>>>>>>>>>>>> Cc: "Ovirt Users"<users@ovirt.org> > >>>>>>>>>>>>>>>>>>>>>>> Sent: Thursday, December 29, 2016 12:47:04 AM > >>>>>>>>>>>>>>>>>>>>>>> Subject: Re: [ovirt-users] Issue with OVN/OVS and > >>>>>>>>>>>>>>>>>>>>>>> mandatory > >>>>>>>>>>>>>>>>>>>>>>> ovirtmgmt > >>>>>>>>>>>>>>>>>>>>>>> network > >>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>> From > >>>>>>>>>>>>>>>>>>>>>>> /usr/libexec/vdsm/hooks/before_device_create/ovirt_provider_ovn_hook > >>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>> (installed by ovirt-provider-ovn-driver rpm): > >>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>> BRIDGE_NAME = 'br-int' > >>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>> Den 2016-12-28 kl. 23:56, skrev Sverker > >>>>>>>>>>>>>>>>>>>>>>> Abrahamsson: > >>>>>>>>>>>>>>>>>>>>>>>> Googling on the message about br-int suggested > >>>>>>>>>>>>>>>>>>>>>>>> adding > >>>>>>>>>>>>>>>>>>>>>>>> that > >>>>>>>>>>>>>>>>>>>>>>>> bridge to > >>>>>>>>>>>>>>>>>>>>>>>> ovs: > >>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>> ovs-vsctl add-br br-int > >>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>> Then the VM is able to boot, but it fails to get > >>>>>>>>>>>>>>>>>>>>>>>> network > >>>>>>>>>>>>>>>>>>>>>>>> connectivity. > >>>>>>>>>>>>>>>>>>>>>>>> Output in /var/log/messages: > >>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>> Dec 28 23:31:35 h2 ovs-vsctl: > >>>>>>>>>>>>>>>>>>>>>>>> ovs|00001|vsctl|INFO|Called as > >>>>>>>>>>>>>>>>>>>>>>>> ovs-vsctl > >>>>>>>>>>>>>>>>>>>>>>>> --timeout=5 -- --if-exists del-port vnet0 -- > >>>>>>>>>>>>>>>>>>>>>>>> add-port > >>>>>>>>>>>>>>>>>>>>>>>> br-int > >>>>>>>>>>>>>>>>>>>>>>>> vnet0 -- > >>>>>>>>>>>>>>>>>>>>>>>> set Interface vnet0 > >>>>>>>>>>>>>>>>>>>>>>>> "external-ids:attached-mac=\"00:1a:4a:16:01:51\"" > >>>>>>>>>>>>>>>>>>>>>>>> -- set Interface vnet0 > >>>>>>>>>>>>>>>>>>>>>>>> "external-ids:iface-id=\"e8853aac-8a75-41b0-8010-e630017dcdd8\"" > >>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>> -- > >>>>>>>>>>>>>>>>>>>>>>>> set Interface vnet0 > >>>>>>>>>>>>>>>>>>>>>>>> "external-ids:vm-id=\"b9440d60-ef5a-4e2b-83cf-081df7c09e6f\"" > >>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>> -- > >>>>>>>>>>>>>>>>>>>>>>>> set > >>>>>>>>>>>>>>>>>>>>>>>> Interface vnet0 external-ids:iface-status=active > >>>>>>>>>>>>>>>>>>>>>>>> Dec 28 23:31:35 h2 kernel: device vnet0 entered > >>>>>>>>>>>>>>>>>>>>>>>> promiscuous > >>>>>>>>>>>>>>>>>>>>>>>> mode > >>>>>>>>>>>>>>>>>>>>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: > >>>>>>>>>>>>>>>>>>>>>>>> COMMAND_FAILED: > >>>>>>>>>>>>>>>>>>>>>>>> '/usr/sbin/ebtables --concurrent -t nat -D > >>>>>>>>>>>>>>>>>>>>>>>> PREROUTING > >>>>>>>>>>>>>>>>>>>>>>>> -i vnet0 > >>>>>>>>>>>>>>>>>>>>>>>> -j > >>>>>>>>>>>>>>>>>>>>>>>> libvirt-J-vnet0' failed: > >>>>>>>>>>>>>>>>>>>>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: > >>>>>>>>>>>>>>>>>>>>>>>> COMMAND_FAILED: > >>>>>>>>>>>>>>>>>>>>>>>> '/usr/sbin/ebtables --concurrent -t nat -D > >>>>>>>>>>>>>>>>>>>>>>>> POSTROUTING -o > >>>>>>>>>>>>>>>>>>>>>>>> vnet0 > >>>>>>>>>>>>>>>>>>>>>>>> -j > >>>>>>>>>>>>>>>>>>>>>>>> libvirt-P-vnet0' failed: > >>>>>>>>>>>>>>>>>>>>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: > >>>>>>>>>>>>>>>>>>>>>>>> COMMAND_FAILED: > >>>>>>>>>>>>>>>>>>>>>>>> '/usr/sbin/ebtables --concurrent -t nat -L > >>>>>>>>>>>>>>>>>>>>>>>> libvirt-J-vnet0' > >>>>>>>>>>>>>>>>>>>>>>>> failed: > >>>>>>>>>>>>>>>>>>>>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: > >>>>>>>>>>>>>>>>>>>>>>>> COMMAND_FAILED: > >>>>>>>>>>>>>>>>>>>>>>>> '/usr/sbin/ebtables --concurrent -t nat -L > >>>>>>>>>>>>>>>>>>>>>>>> libvirt-P-vnet0' > >>>>>>>>>>>>>>>>>>>>>>>> failed: > >>>>>>>>>>>>>>>>>>>>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: > >>>>>>>>>>>>>>>>>>>>>>>> COMMAND_FAILED: > >>>>>>>>>>>>>>>>>>>>>>>> '/usr/sbin/ebtables --concurrent -t nat -F > >>>>>>>>>>>>>>>>>>>>>>>> libvirt-J-vnet0' > >>>>>>>>>>>>>>>>>>>>>>>> failed: > >>>>>>>>>>>>>>>>>>>>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: > >>>>>>>>>>>>>>>>>>>>>>>> COMMAND_FAILED: > >>>>>>>>>>>>>>>>>>>>>>>> '/usr/sbin/ebtables --concurrent -t nat -X > >>>>>>>>>>>>>>>>>>>>>>>> libvirt-J-vnet0' > >>>>>>>>>>>>>>>>>>>>>>>> failed: > >>>>>>>>>>>>>>>>>>>>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: > >>>>>>>>>>>>>>>>>>>>>>>> COMMAND_FAILED: > >>>>>>>>>>>>>>>>>>>>>>>> '/usr/sbin/ebtables --concurrent -t nat -F > >>>>>>>>>>>>>>>>>>>>>>>> libvirt-P-vnet0' > >>>>>>>>>>>>>>>>>>>>>>>> failed: > >>>>>>>>>>>>>>>>>>>>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: > >>>>>>>>>>>>>>>>>>>>>>>> COMMAND_FAILED: > >>>>>>>>>>>>>>>>>>>>>>>> '/usr/sbin/ebtables --concurrent -t nat -X > >>>>>>>>>>>>>>>>>>>>>>>> libvirt-P-vnet0' > >>>>>>>>>>>>>>>>>>>>>>>> failed: > >>>>>>>>>>>>>>>>>>>>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: > >>>>>>>>>>>>>>>>>>>>>>>> COMMAND_FAILED: > >>>>>>>>>>>>>>>>>>>>>>>> '/usr/sbin/ebtables --concurrent -t nat -F > >>>>>>>>>>>>>>>>>>>>>>>> J-vnet0-mac' > >>>>>>>>>>>>>>>>>>>>>>>> failed: > >>>>>>>>>>>>>>>>>>>>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: > >>>>>>>>>>>>>>>>>>>>>>>> COMMAND_FAILED: > >>>>>>>>>>>>>>>>>>>>>>>> '/usr/sbin/ebtables --concurrent -t nat -X > >>>>>>>>>>>>>>>>>>>>>>>> J-vnet0-mac' > >>>>>>>>>>>>>>>>>>>>>>>> failed: > >>>>>>>>>>>>>>>>>>>>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: > >>>>>>>>>>>>>>>>>>>>>>>> COMMAND_FAILED: > >>>>>>>>>>>>>>>>>>>>>>>> '/usr/sbin/ebtables --concurrent -t nat -F > >>>>>>>>>>>>>>>>>>>>>>>> J-vnet0-arp-mac' > >>>>>>>>>>>>>>>>>>>>>>>> failed: > >>>>>>>>>>>>>>>>>>>>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: > >>>>>>>>>>>>>>>>>>>>>>>> COMMAND_FAILED: > >>>>>>>>>>>>>>>>>>>>>>>> '/usr/sbin/ebtables --concurrent -t nat -X > >>>>>>>>>>>>>>>>>>>>>>>> J-vnet0-arp-mac' > >>>>>>>>>>>>>>>>>>>>>>>> failed: > >>>>>>>>>>>>>>>>>>>>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: > >>>>>>>>>>>>>>>>>>>>>>>> COMMAND_FAILED: > >>>>>>>>>>>>>>>>>>>>>>>> '/usr/sbin/iptables -w2 -w -D libvirt-out -m > >>>>>>>>>>>>>>>>>>>>>>>> physdev > >>>>>>>>>>>>>>>>>>>>>>>> --physdev-is-bridged --physdev-out vnet0 -g > >>>>>>>>>>>>>>>>>>>>>>>> FO-vnet0' > >>>>>>>>>>>>>>>>>>>>>>>> failed: > >>>>>>>>>>>>>>>>>>>>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: > >>>>>>>>>>>>>>>>>>>>>>>> COMMAND_FAILED: > >>>>>>>>>>>>>>>>>>>>>>>> '/usr/sbin/iptables -w2 -w -D libvirt-out -m > >>>>>>>>>>>>>>>>>>>>>>>> physdev > >>>>>>>>>>>>>>>>>>>>>>>> --physdev-out > >>>>>>>>>>>>>>>>>>>>>>>> vnet0 -g FO-vnet0' failed: > >>>>>>>>>>>>>>>>>>>>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: > >>>>>>>>>>>>>>>>>>>>>>>> COMMAND_FAILED: > >>>>>>>>>>>>>>>>>>>>>>>> '/usr/sbin/iptables -w2 -w -D libvirt-in -m > >>>>>>>>>>>>>>>>>>>>>>>> physdev > >>>>>>>>>>>>>>>>>>>>>>>> --physdev-in > >>>>>>>>>>>>>>>>>>>>>>>> vnet0 > >>>>>>>>>>>>>>>>>>>>>>>> -g FI-vnet0' failed: > >>>>>>>>>>>>>>>>>>>>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: > >>>>>>>>>>>>>>>>>>>>>>>> COMMAND_FAILED: > >>>>>>>>>>>>>>>>>>>>>>>> '/usr/sbin/iptables -w2 -w -D libvirt-host-in -m > >>>>>>>>>>>>>>>>>>>>>>>> physdev > >>>>>>>>>>>>>>>>>>>>>>>> --physdev-in > >>>>>>>>>>>>>>>>>>>>>>>> vnet0 -g HI-vnet0' failed: > >>>>>>>>>>>>>>>>>>>>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: > >>>>>>>>>>>>>>>>>>>>>>>> COMMAND_FAILED: > >>>>>>>>>>>>>>>>>>>>>>>> '/usr/sbin/iptables -w2 -w -F FO-vnet0' failed: > >>>>>>>>>>>>>>>>>>>>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: > >>>>>>>>>>>>>>>>>>>>>>>> COMMAND_FAILED: > >>>>>>>>>>>>>>>>>>>>>>>> '/usr/sbin/iptables -w2 -w -X FO-vnet0' failed: > >>>>>>>>>>>>>>>>>>>>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: > >>>>>>>>>>>>>>>>>>>>>>>> COMMAND_FAILED: > >>>>>>>>>>>>>>>>>>>>>>>> '/usr/sbin/iptables -w2 -w -F FI-vnet0' failed: > >>>>>>>>>>>>>>>>>>>>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: > >>>>>>>>>>>>>>>>>>>>>>>> COMMAND_FAILED: > >>>>>>>>>>>>>>>>>>>>>>>> '/usr/sbin/iptables -w2 -w -X FI-vnet0' failed: > >>>>>>>>>>>>>>>>>>>>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: > >>>>>>>>>>>>>>>>>>>>>>>> COMMAND_FAILED: > >>>>>>>>>>>>>>>>>>>>>>>> '/usr/sbin/iptables -w2 -w -F HI-vnet0' failed: > >>>>>>>>>>>>>>>>>>>>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: > >>>>>>>>>>>>>>>>>>>>>>>> COMMAND_FAILED: > >>>>>>>>>>>>>>>>>>>>>>>> '/usr/sbin/iptables -w2 -w -X HI-vnet0' failed: > >>>>>>>>>>>>>>>>>>>>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: > >>>>>>>>>>>>>>>>>>>>>>>> COMMAND_FAILED: > >>>>>>>>>>>>>>>>>>>>>>>> '/usr/sbin/iptables -w2 -w -E FP-vnet0 FO-vnet0' > >>>>>>>>>>>>>>>>>>>>>>>> failed: > >>>>>>>>>>>>>>>>>>>>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: > >>>>>>>>>>>>>>>>>>>>>>>> COMMAND_FAILED: > >>>>>>>>>>>>>>>>>>>>>>>> '/usr/sbin/iptables -w2 -w -E FJ-vnet0 FI-vnet0' > >>>>>>>>>>>>>>>>>>>>>>>> failed: > >>>>>>>>>>>>>>>>>>>>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: > >>>>>>>>>>>>>>>>>>>>>>>> COMMAND_FAILED: > >>>>>>>>>>>>>>>>>>>>>>>> '/usr/sbin/iptables -w2 -w -E HJ-vnet0 HI-vnet0' > >>>>>>>>>>>>>>>>>>>>>>>> failed: > >>>>>>>>>>>>>>>>>>>>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: > >>>>>>>>>>>>>>>>>>>>>>>> COMMAND_FAILED: > >>>>>>>>>>>>>>>>>>>>>>>> '/usr/sbin/ip6tables -w2 -w -D libvirt-out -m > >>>>>>>>>>>>>>>>>>>>>>>> physdev > >>>>>>>>>>>>>>>>>>>>>>>> --physdev-is-bridged --physdev-out vnet0 -g > >>>>>>>>>>>>>>>>>>>>>>>> FO-vnet0' > >>>>>>>>>>>>>>>>>>>>>>>> failed: > >>>>>>>>>>>>>>>>>>>>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: > >>>>>>>>>>>>>>>>>>>>>>>> COMMAND_FAILED: > >>>>>>>>>>>>>>>>>>>>>>>> '/usr/sbin/ip6tables -w2 -w -D libvirt-out -m > >>>>>>>>>>>>>>>>>>>>>>>> physdev > >>>>>>>>>>>>>>>>>>>>>>>> --physdev-out > >>>>>>>>>>>>>>>>>>>>>>>> vnet0 -g FO-vnet0' failed: > >>>>>>>>>>>>>>>>>>>>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: > >>>>>>>>>>>>>>>>>>>>>>>> COMMAND_FAILED: > >>>>>>>>>>>>>>>>>>>>>>>> '/usr/sbin/ip6tables -w2 -w -D libvirt-in -m > >>>>>>>>>>>>>>>>>>>>>>>> physdev > >>>>>>>>>>>>>>>>>>>>>>>> --physdev-in > >>>>>>>>>>>>>>>>>>>>>>>> vnet0 -g FI-vnet0' failed: > >>>>>>>>>>>>>>>>>>>>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: > >>>>>>>>>>>>>>>>>>>>>>>> COMMAND_FAILED: > >>>>>>>>>>>>>>>>>>>>>>>> '/usr/sbin/ip6tables -w2 -w -D libvirt-host-in -m > >>>>>>>>>>>>>>>>>>>>>>>> physdev > >>>>>>>>>>>>>>>>>>>>>>>> --physdev-in > >>>>>>>>>>>>>>>>>>>>>>>> vnet0 -g HI-vnet0' failed: > >>>>>>>>>>>>>>>>>>>>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: > >>>>>>>>>>>>>>>>>>>>>>>> COMMAND_FAILED: > >>>>>>>>>>>>>>>>>>>>>>>> '/usr/sbin/ip6tables -w2 -w -F FO-vnet0' failed: > >>>>>>>>>>>>>>>>>>>>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: > >>>>>>>>>>>>>>>>>>>>>>>> COMMAND_FAILED: > >>>>>>>>>>>>>>>>>>>>>>>> '/usr/sbin/ip6tables -w2 -w -X FO-vnet0' failed: > >>>>>>>>>>>>>>>>>>>>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: > >>>>>>>>>>>>>>>>>>>>>>>> COMMAND_FAILED: > >>>>>>>>>>>>>>>>>>>>>>>> '/usr/sbin/ip6tables -w2 -w -F FI-vnet0' failed: > >>>>>>>>>>>>>>>>>>>>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: > >>>>>>>>>>>>>>>>>>>>>>>> COMMAND_FAILED: > >>>>>>>>>>>>>>>>>>>>>>>> '/usr/sbin/ip6tables -w2 -w -X FI-vnet0' failed: > >>>>>>>>>>>>>>>>>>>>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: > >>>>>>>>>>>>>>>>>>>>>>>> COMMAND_FAILED: > >>>>>>>>>>>>>>>>>>>>>>>> '/usr/sbin/ip6tables -w2 -w -F HI-vnet0' failed: > >>>>>>>>>>>>>>>>>>>>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: > >>>>>>>>>>>>>>>>>>>>>>>> COMMAND_FAILED: > >>>>>>>>>>>>>>>>>>>>>>>> '/usr/sbin/ip6tables -w2 -w -X HI-vnet0' failed: > >>>>>>>>>>>>>>>>>>>>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: > >>>>>>>>>>>>>>>>>>>>>>>> COMMAND_FAILED: > >>>>>>>>>>>>>>>>>>>>>>>> '/usr/sbin/ip6tables -w2 -w -E FP-vnet0 FO-vnet0' > >>>>>>>>>>>>>>>>>>>>>>>> failed: > >>>>>>>>>>>>>>>>>>>>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: > >>>>>>>>>>>>>>>>>>>>>>>> COMMAND_FAILED: > >>>>>>>>>>>>>>>>>>>>>>>> '/usr/sbin/ip6tables -w2 -w -E FJ-vnet0 FI-vnet0' > >>>>>>>>>>>>>>>>>>>>>>>> failed: > >>>>>>>>>>>>>>>>>>>>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: > >>>>>>>>>>>>>>>>>>>>>>>> COMMAND_FAILED: > >>>>>>>>>>>>>>>>>>>>>>>> '/usr/sbin/ip6tables -w2 -w -E HJ-vnet0 HI-vnet0' > >>>>>>>>>>>>>>>>>>>>>>>> failed: > >>>>>>>>>>>>>>>>>>>>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: > >>>>>>>>>>>>>>>>>>>>>>>> COMMAND_FAILED: > >>>>>>>>>>>>>>>>>>>>>>>> '/usr/sbin/ebtables --concurrent -t nat -D > >>>>>>>>>>>>>>>>>>>>>>>> PREROUTING > >>>>>>>>>>>>>>>>>>>>>>>> -i vnet0 > >>>>>>>>>>>>>>>>>>>>>>>> -j > >>>>>>>>>>>>>>>>>>>>>>>> libvirt-I-vnet0' failed: > >>>>>>>>>>>>>>>>>>>>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: > >>>>>>>>>>>>>>>>>>>>>>>> COMMAND_FAILED: > >>>>>>>>>>>>>>>>>>>>>>>> '/usr/sbin/ebtables --concurrent -t nat -D > >>>>>>>>>>>>>>>>>>>>>>>> POSTROUTING -o > >>>>>>>>>>>>>>>>>>>>>>>> vnet0 > >>>>>>>>>>>>>>>>>>>>>>>> -j > >>>>>>>>>>>>>>>>>>>>>>>> libvirt-O-vnet0' failed: > >>>>>>>>>>>>>>>>>>>>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: > >>>>>>>>>>>>>>>>>>>>>>>> COMMAND_FAILED: > >>>>>>>>>>>>>>>>>>>>>>>> '/usr/sbin/ebtables --concurrent -t nat -L > >>>>>>>>>>>>>>>>>>>>>>>> libvirt-I-vnet0' > >>>>>>>>>>>>>>>>>>>>>>>> failed: > >>>>>>>>>>>>>>>>>>>>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: > >>>>>>>>>>>>>>>>>>>>>>>> COMMAND_FAILED: > >>>>>>>>>>>>>>>>>>>>>>>> '/usr/sbin/ebtables --concurrent -t nat -L > >>>>>>>>>>>>>>>>>>>>>>>> libvirt-O-vnet0' > >>>>>>>>>>>>>>>>>>>>>>>> failed: > >>>>>>>>>>>>>>>>>>>>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: > >>>>>>>>>>>>>>>>>>>>>>>> COMMAND_FAILED: > >>>>>>>>>>>>>>>>>>>>>>>> '/usr/sbin/ebtables --concurrent -t nat -F > >>>>>>>>>>>>>>>>>>>>>>>> libvirt-I-vnet0' > >>>>>>>>>>>>>>>>>>>>>>>> failed: > >>>>>>>>>>>>>>>>>>>>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: > >>>>>>>>>>>>>>>>>>>>>>>> COMMAND_FAILED: > >>>>>>>>>>>>>>>>>>>>>>>> '/usr/sbin/ebtables --concurrent -t nat -X > >>>>>>>>>>>>>>>>>>>>>>>> libvirt-I-vnet0' > >>>>>>>>>>>>>>>>>>>>>>>> failed: > >>>>>>>>>>>>>>>>>>>>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: > >>>>>>>>>>>>>>>>>>>>>>>> COMMAND_FAILED: > >>>>>>>>>>>>>>>>>>>>>>>> '/usr/sbin/ebtables --concurrent -t nat -F > >>>>>>>>>>>>>>>>>>>>>>>> libvirt-O-vnet0' > >>>>>>>>>>>>>>>>>>>>>>>> failed: > >>>>>>>>>>>>>>>>>>>>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: > >>>>>>>>>>>>>>>>>>>>>>>> COMMAND_FAILED: > >>>>>>>>>>>>>>>>>>>>>>>> '/usr/sbin/ebtables --concurrent -t nat -X > >>>>>>>>>>>>>>>>>>>>>>>> libvirt-O-vnet0' > >>>>>>>>>>>>>>>>>>>>>>>> failed: > >>>>>>>>>>>>>>>>>>>>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: > >>>>>>>>>>>>>>>>>>>>>>>> COMMAND_FAILED: > >>>>>>>>>>>>>>>>>>>>>>>> '/usr/sbin/ebtables --concurrent -t nat -L > >>>>>>>>>>>>>>>>>>>>>>>> libvirt-P-vnet0' > >>>>>>>>>>>>>>>>>>>>>>>> failed: > >>>>>>>>>>>>>>>>>>>>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: > >>>>>>>>>>>>>>>>>>>>>>>> COMMAND_FAILED: > >>>>>>>>>>>>>>>>>>>>>>>> '/usr/sbin/ebtables --concurrent -t nat -E > >>>>>>>>>>>>>>>>>>>>>>>> libvirt-P-vnet0 > >>>>>>>>>>>>>>>>>>>>>>>> libvirt-O-vnet0' failed: > >>>>>>>>>>>>>>>>>>>>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: > >>>>>>>>>>>>>>>>>>>>>>>> COMMAND_FAILED: > >>>>>>>>>>>>>>>>>>>>>>>> '/usr/sbin/ebtables --concurrent -t nat -F > >>>>>>>>>>>>>>>>>>>>>>>> I-vnet0-mac' > >>>>>>>>>>>>>>>>>>>>>>>> failed: > >>>>>>>>>>>>>>>>>>>>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: > >>>>>>>>>>>>>>>>>>>>>>>> COMMAND_FAILED: > >>>>>>>>>>>>>>>>>>>>>>>> '/usr/sbin/ebtables --concurrent -t nat -X > >>>>>>>>>>>>>>>>>>>>>>>> I-vnet0-mac' > >>>>>>>>>>>>>>>>>>>>>>>> failed: > >>>>>>>>>>>>>>>>>>>>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: > >>>>>>>>>>>>>>>>>>>>>>>> COMMAND_FAILED: > >>>>>>>>>>>>>>>>>>>>>>>> '/usr/sbin/ebtables --concurrent -t nat -F > >>>>>>>>>>>>>>>>>>>>>>>> I-vnet0-arp-mac' > >>>>>>>>>>>>>>>>>>>>>>>> failed: > >>>>>>>>>>>>>>>>>>>>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: > >>>>>>>>>>>>>>>>>>>>>>>> COMMAND_FAILED: > >>>>>>>>>>>>>>>>>>>>>>>> '/usr/sbin/ebtables --concurrent -t nat -X > >>>>>>>>>>>>>>>>>>>>>>>> I-vnet0-arp-mac' > >>>>>>>>>>>>>>>>>>>>>>>> failed: > >>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>> [root@h2 etc]# ovs-vsctl show > >>>>>>>>>>>>>>>>>>>>>>>> ebb6aede-cbbc-4f4f-a88a-a9cd72b2bd23 > >>>>>>>>>>>>>>>>>>>>>>>> Bridge ovirtbridge > >>>>>>>>>>>>>>>>>>>>>>>> Port "ovirtport0" > >>>>>>>>>>>>>>>>>>>>>>>> Interface "ovirtport0" > >>>>>>>>>>>>>>>>>>>>>>>> type: internal > >>>>>>>>>>>>>>>>>>>>>>>> Port ovirtbridge > >>>>>>>>>>>>>>>>>>>>>>>> Interface ovirtbridge > >>>>>>>>>>>>>>>>>>>>>>>> type: internal > >>>>>>>>>>>>>>>>>>>>>>>> Bridge "ovsbridge0" > >>>>>>>>>>>>>>>>>>>>>>>> Port "ovsbridge0" > >>>>>>>>>>>>>>>>>>>>>>>> Interface "ovsbridge0" > >>>>>>>>>>>>>>>>>>>>>>>> type: internal > >>>>>>>>>>>>>>>>>>>>>>>> Port "eth0" > >>>>>>>>>>>>>>>>>>>>>>>> Interface "eth0" > >>>>>>>>>>>>>>>>>>>>>>>> Bridge br-int > >>>>>>>>>>>>>>>>>>>>>>>> Port br-int > >>>>>>>>>>>>>>>>>>>>>>>> Interface br-int > >>>>>>>>>>>>>>>>>>>>>>>> type: internal > >>>>>>>>>>>>>>>>>>>>>>>> Port "vnet0" > >>>>>>>>>>>>>>>>>>>>>>>> Interface "vnet0" > >>>>>>>>>>>>>>>>>>>>>>>> ovs_version: "2.6.90" > >>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>> Searching through the code it appears that br-int > >>>>>>>>>>>>>>>>>>>>>>>> comes from > >>>>>>>>>>>>>>>>>>>>>>>> neutron-openvswitch plugin ?? > >>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>> [root@h2 share]# rpm -qf > >>>>>>>>>>>>>>>>>>>>>>>> /usr/share/otopi/plugins/ovirt-host-deploy/openstack/neutron_openvswitch.py > >>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>> ovirt-host-deploy-1.6.0-0.0.master.20161215101008.gitb76ad50.el7.centos.noarch > >>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>> /Sverker > >>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>> Den 2016-12-28 kl. 23:24, skrev Sverker > >>>>>>>>>>>>>>>>>>>>>>>> Abrahamsson: > >>>>>>>>>>>>>>>>>>>>>>>>> In addition I had to add an alias to modprobe: > >>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>> [root@h2 modprobe.d]# cat dummy.conf > >>>>>>>>>>>>>>>>>>>>>>>>> alias dummy0 dummy > >>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>> Den 2016-12-28 kl. 23:03, skrev Sverker > >>>>>>>>>>>>>>>>>>>>>>>>> Abrahamsson: > >>>>>>>>>>>>>>>>>>>>>>>>>> Hi > >>>>>>>>>>>>>>>>>>>>>>>>>> I first tried to set device name to dummy_0, but > >>>>>>>>>>>>>>>>>>>>>>>>>> then ifup > >>>>>>>>>>>>>>>>>>>>>>>>>> did > >>>>>>>>>>>>>>>>>>>>>>>>>> not > >>>>>>>>>>>>>>>>>>>>>>>>>> succeed in creating the device unless I first > >>>>>>>>>>>>>>>>>>>>>>>>>> did > >>>>>>>>>>>>>>>>>>>>>>>>>> 'ip link > >>>>>>>>>>>>>>>>>>>>>>>>>> add > >>>>>>>>>>>>>>>>>>>>>>>>>> dummy_0 type dummy' but then it would not > >>>>>>>>>>>>>>>>>>>>>>>>>> suceed to > >>>>>>>>>>>>>>>>>>>>>>>>>> establish > >>>>>>>>>>>>>>>>>>>>>>>>>> the if > >>>>>>>>>>>>>>>>>>>>>>>>>> on reboot. > >>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>> Setting fake_nics = dummy0 would not work > >>>>>>>>>>>>>>>>>>>>>>>>>> neither, > >>>>>>>>>>>>>>>>>>>>>>>>>> but this > >>>>>>>>>>>>>>>>>>>>>>>>>> works: > >>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>> fake_nics = dummy* > >>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>> The engine is now able to find the if and assign > >>>>>>>>>>>>>>>>>>>>>>>>>> bridge > >>>>>>>>>>>>>>>>>>>>>>>>>> ovirtmgmt to > >>>>>>>>>>>>>>>>>>>>>>>>>> it. > >>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>> However, I then run into the next issue when > >>>>>>>>>>>>>>>>>>>>>>>>>> starting a VM: > >>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>> 2016-12-28 22:28:23,897 ERROR > >>>>>>>>>>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] > >>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>> (ForkJoinPool-1-worker-2) [] Correlation ID: > >>>>>>>>>>>>>>>>>>>>>>>>>> null, > >>>>>>>>>>>>>>>>>>>>>>>>>> Call > >>>>>>>>>>>>>>>>>>>>>>>>>> Stack: > >>>>>>>>>>>>>>>>>>>>>>>>>> null, > >>>>>>>>>>>>>>>>>>>>>>>>>> Custom Event ID: -1, Message: VM CentOS7 is down > >>>>>>>>>>>>>>>>>>>>>>>>>> with error. > >>>>>>>>>>>>>>>>>>>>>>>>>> Exit > >>>>>>>>>>>>>>>>>>>>>>>>>> message: Cannot get interface MTU on > >>>>>>>>>>>>>>>>>>>>>>>>>> 'br-int': No > >>>>>>>>>>>>>>>>>>>>>>>>>> such > >>>>>>>>>>>>>>>>>>>>>>>>>> device. > >>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>> This VM has a nic on ovirtbridge, which comes > >>>>>>>>>>>>>>>>>>>>>>>>>> from > >>>>>>>>>>>>>>>>>>>>>>>>>> the OVN > >>>>>>>>>>>>>>>>>>>>>>>>>> provider. > >>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>> /Sverker > >>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>> Den 2016-12-28 kl. 14:38, skrev Marcin Mirecki: > >>>>>>>>>>>>>>>>>>>>>>>>>>> Sverker, > >>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>> Can you try adding a vnic named veth_* or > >>>>>>>>>>>>>>>>>>>>>>>>>>> dummy_*, > >>>>>>>>>>>>>>>>>>>>>>>>>>> (or alternatively add the name of the vnic to > >>>>>>>>>>>>>>>>>>>>>>>>>>> vdsm.config fake_nics), and setup the > >>>>>>>>>>>>>>>>>>>>>>>>>>> management > >>>>>>>>>>>>>>>>>>>>>>>>>>> network using this vnic? > >>>>>>>>>>>>>>>>>>>>>>>>>>> I suppose adding the vnic you use for > >>>>>>>>>>>>>>>>>>>>>>>>>>> connecting > >>>>>>>>>>>>>>>>>>>>>>>>>>> to the engine to fake_nics should make it > >>>>>>>>>>>>>>>>>>>>>>>>>>> visible > >>>>>>>>>>>>>>>>>>>>>>>>>>> to the engine, and you should be able to use > >>>>>>>>>>>>>>>>>>>>>>>>>>> it for > >>>>>>>>>>>>>>>>>>>>>>>>>>> the setup. > >>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>> Marcin > >>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>> ----- Original Message ----- > >>>>>>>>>>>>>>>>>>>>>>>>>>>> From: "Marcin Mirecki"<mmire...@redhat.com> > >>>>>>>>>>>>>>>>>>>>>>>>>>>> To: "Sverker > >>>>>>>>>>>>>>>>>>>>>>>>>>>> Abrahamsson"<sver...@abrahamsson.com> > >>>>>>>>>>>>>>>>>>>>>>>>>>>> Cc: "Ovirt Users"<users@ovirt.org> > >>>>>>>>>>>>>>>>>>>>>>>>>>>> Sent: Wednesday, December 28, 2016 12:06:26 PM > >>>>>>>>>>>>>>>>>>>>>>>>>>>> Subject: Re: [ovirt-users] Issue with > >>>>>>>>>>>>>>>>>>>>>>>>>>>> OVN/OVS and > >>>>>>>>>>>>>>>>>>>>>>>>>>>> mandatory > >>>>>>>>>>>>>>>>>>>>>>>>>>>> ovirtmgmt network > >>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>> I have an internal OVS bridge called > >>>>>>>>>>>>>>>>>>>>>>>>>>>>> ovirtbridge > >>>>>>>>>>>>>>>>>>>>>>>>>>>>> which > >>>>>>>>>>>>>>>>>>>>>>>>>>>>> has > >>>>>>>>>>>>>>>>>>>>>>>>>>>>> a port > >>>>>>>>>>>>>>>>>>>>>>>>>>>>> with > >>>>>>>>>>>>>>>>>>>>>>>>>>>>> IP address, but in the host network settings > >>>>>>>>>>>>>>>>>>>>>>>>>>>>> that port is > >>>>>>>>>>>>>>>>>>>>>>>>>>>>> not > >>>>>>>>>>>>>>>>>>>>>>>>>>>>> visible. > >>>>>>>>>>>>>>>>>>>>>>>>>>>> I just verified and unfortunately the virtual > >>>>>>>>>>>>>>>>>>>>>>>>>>>> ports are > >>>>>>>>>>>>>>>>>>>>>>>>>>>> not > >>>>>>>>>>>>>>>>>>>>>>>>>>>> visible in engine > >>>>>>>>>>>>>>>>>>>>>>>>>>>> to assign a network to :( > >>>>>>>>>>>>>>>>>>>>>>>>>>>> I'm afraid that the engine is not ready for > >>>>>>>>>>>>>>>>>>>>>>>>>>>> such a > >>>>>>>>>>>>>>>>>>>>>>>>>>>> scenario > >>>>>>>>>>>>>>>>>>>>>>>>>>>> (even > >>>>>>>>>>>>>>>>>>>>>>>>>>>> if it > >>>>>>>>>>>>>>>>>>>>>>>>>>>> works). > >>>>>>>>>>>>>>>>>>>>>>>>>>>> Please give me some time to look for a > >>>>>>>>>>>>>>>>>>>>>>>>>>>> solution. > >>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>> ----- Original Message ----- > >>>>>>>>>>>>>>>>>>>>>>>>>>>>> From: "Sverker > >>>>>>>>>>>>>>>>>>>>>>>>>>>>> Abrahamsson"<sver...@abrahamsson.com> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>> To: "Marcin Mirecki"<mmire...@redhat.com> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>> Cc: "Ovirt Users"<users@ovirt.org> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>> Sent: Wednesday, December 28, 2016 > >>>>>>>>>>>>>>>>>>>>>>>>>>>>> 11:48:24 AM > >>>>>>>>>>>>>>>>>>>>>>>>>>>>> Subject: Re: [ovirt-users] Issue with > >>>>>>>>>>>>>>>>>>>>>>>>>>>>> OVN/OVS and > >>>>>>>>>>>>>>>>>>>>>>>>>>>>> mandatory > >>>>>>>>>>>>>>>>>>>>>>>>>>>>> ovirtmgmt > >>>>>>>>>>>>>>>>>>>>>>>>>>>>> network > >>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>> Hi Marcin > >>>>>>>>>>>>>>>>>>>>>>>>>>>>> Yes, that is my issue. I don't want to let > >>>>>>>>>>>>>>>>>>>>>>>>>>>>> ovirt/vdsm see > >>>>>>>>>>>>>>>>>>>>>>>>>>>>> eth0 > >>>>>>>>>>>>>>>>>>>>>>>>>>>>> nor > >>>>>>>>>>>>>>>>>>>>>>>>>>>>> ovsbridge0 since as soon as it sees them it > >>>>>>>>>>>>>>>>>>>>>>>>>>>>> messes up the > >>>>>>>>>>>>>>>>>>>>>>>>>>>>> network > >>>>>>>>>>>>>>>>>>>>>>>>>>>>> config > >>>>>>>>>>>>>>>>>>>>>>>>>>>>> so that the host will be unreachable. > >>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>> I have an internal OVS bridge called > >>>>>>>>>>>>>>>>>>>>>>>>>>>>> ovirtbridge > >>>>>>>>>>>>>>>>>>>>>>>>>>>>> which > >>>>>>>>>>>>>>>>>>>>>>>>>>>>> has > >>>>>>>>>>>>>>>>>>>>>>>>>>>>> a port > >>>>>>>>>>>>>>>>>>>>>>>>>>>>> with > >>>>>>>>>>>>>>>>>>>>>>>>>>>>> IP address, but in the host network settings > >>>>>>>>>>>>>>>>>>>>>>>>>>>>> that port is > >>>>>>>>>>>>>>>>>>>>>>>>>>>>> not > >>>>>>>>>>>>>>>>>>>>>>>>>>>>> visible. > >>>>>>>>>>>>>>>>>>>>>>>>>>>>> It doesn't help to name it ovirtmgmt. > >>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>> The engine is able to communicate with the > >>>>>>>>>>>>>>>>>>>>>>>>>>>>> host > >>>>>>>>>>>>>>>>>>>>>>>>>>>>> on the ip > >>>>>>>>>>>>>>>>>>>>>>>>>>>>> it has > >>>>>>>>>>>>>>>>>>>>>>>>>>>>> been > >>>>>>>>>>>>>>>>>>>>>>>>>>>>> given, it's just that it believes that it > >>>>>>>>>>>>>>>>>>>>>>>>>>>>> HAS to > >>>>>>>>>>>>>>>>>>>>>>>>>>>>> have a > >>>>>>>>>>>>>>>>>>>>>>>>>>>>> ovirtmgmt > >>>>>>>>>>>>>>>>>>>>>>>>>>>>> network which can't be on OVN. > >>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>> /Sverker > >>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>> Den 2016-12-28 kl. 10:45, skrev Marcin > >>>>>>>>>>>>>>>>>>>>>>>>>>>>> Mirecki: > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> Hi Sverker, > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> The management network is mandatory on each > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> host. It's > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> used by > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> the > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> engine to communicate with the host. > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> Looking at your description and the > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> exception > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> it looks > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> like it > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> is > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> missing. > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> The error is caused by not having any > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> network > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> for the > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> host > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> (network list retrieved in > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> InterfaceDaoImpl.getHostNetworksByCluster - > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> which > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> gets all the networks on nics for a host > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> from > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> vds_interface > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> table in the > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> DB). > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> Could you maybe create a virtual nic > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> connected to > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> ovsbridge0 (as > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> I > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> understand you > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> have no physical nic available) and use this > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> for the > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> management > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> network? > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> I then create a bridge for use with > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> ovirt, with > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> a > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> private > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> address. > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> I'm not quite sure I understand. Is this yet > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> another > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> bridge > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> connected to > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> ovsbridge0? > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> You could also attach the vnic for the > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> management > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> network > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> here > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> if need > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> be. > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> Please keep in mind that OVN has no use in > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> setting up > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> the > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> management > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> network. > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> The OVN provider can only handle external > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> networks, > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> which > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> can > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> not be used > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> for a > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> management network. > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> Marcin > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>> ----- Original Message ----- > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> From: "Sverker > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> Abrahamsson"<sver...@abrahamsson.com> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> To:users@ovirt.org > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> Sent: Wednesday, December 28, 2016 > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> 12:39:59 AM > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> Subject: [ovirt-users] Issue with > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> OVN/OVS and > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> mandatory > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> ovirtmgmt > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> network > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> Hi > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> For long time I've been looking for proper > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> support in > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> ovirt for > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> Open > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> vSwitch > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> so I'm happy that it is moving in the right > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> direction. > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> However, > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> there > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> seems > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> to still be a dependency on a ovirtmgmt > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> bridge > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> and I'm > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> unable > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> to move > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> that > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> to the OVN provider. > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> The hosting center where I rent hw > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> instances > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> has a bit > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> special > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> network > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> setup, > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> so I have one physical network port with > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> a /32 > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> netmask > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> and > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> point-to-point > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> config to router. The physical port I > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> connect > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> to a ovs > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> bridge > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> which has > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> the > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> public ip. Since ovirt always messes up the > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> network > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> config when > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> I've > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> tried > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> to let it have access to the network config > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> for the > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> physical > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> port, I've > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> set > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> eht0 and ovsbridge0 as hidden in vdsm.conf. > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> I then create a bridge for use with > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> ovirt, with > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> a > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> private > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> address. With > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> the > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> OVN provider I am now able to import these > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> into the > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> engine and > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> it looks > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> good. When creating a VM I can select > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> that it > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> will have > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> a > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> vNic > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> on my OVS > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> bridge. > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> However, I can't start the VM as an > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> exception > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> is thrown > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> in the > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> log: > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> 2016-12-28 00:13:33,350 ERROR > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.bll.RunVmCommand] > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> (default task-5) [3c882d53] Error during > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> ValidateFailure.: > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> java.lang.NullPointerException > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> at > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> org.ovirt.engine.core.bll.scheduling.policyunits.NetworkPolicyUnit.validateRequiredNetworksAvailable(NetworkPolicyUnit.java:140) > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> [bll.jar:] > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> at > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> org.ovirt.engine.core.bll.scheduling.policyunits.NetworkPolicyUnit.filter(NetworkPolicyUnit.java:69) > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> [bll.jar:] > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> at > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> org.ovirt.engine.core.bll.scheduling.SchedulingManager.runInternalFilters(SchedulingManager.java:597) > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> [bll.jar:] > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> at > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> org.ovirt.engine.core.bll.scheduling.SchedulingManager.runFilters(SchedulingManager.java:564) > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> [bll.jar:] > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> at > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> org.ovirt.engine.core.bll.scheduling.SchedulingManager.canSchedule(SchedulingManager.java:494) > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> [bll.jar:] > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> at > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> org.ovirt.engine.core.bll.validator.RunVmValidator.canRunVm(RunVmValidator.java:133) > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> [bll.jar:] > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> at > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> org.ovirt.engine.core.bll.RunVmCommand.validate(RunVmCommand.java:940) > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> [bll.jar:] > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> at > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> org.ovirt.engine.core.bll.CommandBase.internalValidate(CommandBase.java:886) > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> [bll.jar:] > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> at > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> org.ovirt.engine.core.bll.CommandBase.validateOnly(CommandBase.java:366) > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> [bll.jar:] > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> at > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> org.ovirt.engine.core.bll.PrevalidatingMultipleActionsRunner.canRunActions(PrevalidatingMultipleActionsRunner.java:113) > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> [bll.jar:] > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> at > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> org.ovirt.engine.core.bll.PrevalidatingMultipleActionsRunner.invokeCommands(PrevalidatingMultipleActionsRunner.java:99) > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> [bll.jar:] > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> at > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> org.ovirt.engine.core.bll.PrevalidatingMultipleActionsRunner.execute(PrevalidatingMultipleActionsRunner.java:76) > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> [bll.jar:] > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> at > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> org.ovirt.engine.core.bll.Backend.runMultipleActionsImpl(Backend.java:613) > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> [bll.jar:] > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> at > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> org.ovirt.engine.core.bll.Backend.runMultipleActions(Backend.java:583) > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> [bll.jar:] > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> Looking at that section of code where the > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> exception is > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> thrown, > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> I see > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> that > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> it > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> iterates over host networks to find > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> required > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> networks, > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> which I > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> assume is > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> ovirtmgmt. In the host network setup > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> dialog I > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> don't see > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> any > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> networks at > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> all > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> but it lists ovirtmgmt as required. It also > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> list the > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> OVN > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> networks but > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> these > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> can't be statically assigned as they are > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> added > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> dynamically when > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> needed, > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> which is fine. > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> I believe that I either need to remove > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> ovirtmgmt > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> network > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> or > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> configure > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> that > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> it > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> is provided by the OVN provider, but > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> neither is > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> possible. > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> Preferably it > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> shouldn't be hardcoded which network is > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> management and > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> mandatory but be > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> possible to configure. > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> /Sverker > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> Den 2016-12-27 kl. 17:10, skrev Marcin > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> Mirecki: > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>> _______________________________________________ > >>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>>> Users mailing list > >>>>>>>>>>>>>>>>>>>>>>>>>>>> Users@ovirt.org > >>>>>>>>>>>>>>>>>>>>>>>>>>>> http://lists.ovirt.org/mailman/listinfo/users > >>>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>> _______________________________________________ > >>>>>>>>>>>>>>>>>>>>>>>>>> Users mailing list > >>>>>>>>>>>>>>>>>>>>>>>>>> Users@ovirt.org > >>>>>>>>>>>>>>>>>>>>>>>>>> http://lists.ovirt.org/mailman/listinfo/users > >>>>>>>>>>>>>>>>>>>>>>>>> _______________________________________________ > >>>>>>>>>>>>>>>>>>>>>>>>> Users mailing list > >>>>>>>>>>>>>>>>>>>>>>>>> Users@ovirt.org > >>>>>>>>>>>>>>>>>>>>>>>>> http://lists.ovirt.org/mailman/listinfo/users > >>>>>>>>>>>>>>>>>>>>>>>> _______________________________________________ > >>>>>>>>>>>>>>>>>>>>>>>> Users mailing list > >>>>>>>>>>>>>>>>>>>>>>>> Users@ovirt.org > >>>>>>>>>>>>>>>>>>>>>>>> http://lists.ovirt.org/mailman/listinfo/users > >>>>>>>>>>>>>>>>>>>>>> _______________________________________________ > >>>>>>>>>>>>>>>>>>>>>> Users mailing list > >>>>>>>>>>>>>>>>>>>>>> Users@ovirt.org > >>>>>>>>>>>>>>>>>>>>>> http://lists.ovirt.org/mailman/listinfo/users > >>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> _______________________________________________ > >>>>>>>>>>>>>>>>> Users mailing list > >>>>>>>>>>>>>>>>> Users@ovirt.org > >>>>>>>>>>>>>>>>> http://lists.ovirt.org/mailman/listinfo/users > >>>>>>>>>>>>>>>> _______________________________________________ > >>>>>>>>>>>>>>>> Users mailing list > >>>>>>>>>>>>>>>> Users@ovirt.org > >>>>>>>>>>>>>>>> http://lists.ovirt.org/mailman/listinfo/users > >>>>>>>>>>>>>>> _______________________________________________ > >>>>>>>>>>>>>>> Users mailing list > >>>>>>>>>>>>>>> Users@ovirt.org > >>>>>>>>>>>>>>> http://lists.ovirt.org/mailman/listinfo/users > >>>>>>>> _______________________________________________ > >>>>>>>> Users mailing list > >>>>>>>> Users@ovirt.org > >>>>>>>> http://lists.ovirt.org/mailman/listinfo/users > >>>>>>> _______________________________________________ > >>>>>>> Users mailing list > >>>>>>> Users@ovirt.org > >>>>>>> http://lists.ovirt.org/mailman/listinfo/users > >>>> > >>>> > >>>> _______________________________________________ > >>>> Users mailing list > >>>> Users@ovirt.org > >>>> http://lists.ovirt.org/mailman/listinfo/users > >>> > > > > > > _______________________________________________ > > Users mailing list > > Users@ovirt.org > > http://lists.ovirt.org/mailman/listinfo/users > > _______________________________________________ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users