Public bug reported:

Hello guys.

As I found in 2021 there was a commit to OVS [since OVS v2.16 or 2.15 with that 
backport] that changed behavior during OVN DBs snapshoting. 
Now before the leader creates a snapshot it will transfer leadership to another 
node. 

We've run tests with rally and tempest and looks like there is a problem now 
when there is interaction between nova and neutron. 
For example, simple rally test like 'create {network,router,subnet} -> add 
interface to router' looks okay even with 256 concurrent same tests/threads. 
But something like 'neutron.create_subnet -> nova.boot_server -> 
nova.attach_interface' will fail in time when transfer leadership happens. 
Since it happens to often [ each 10m + rand(10) ] we will get a lot of errors. 

This problem can be observed on all versions where OVS 2.16 [or
backport] or higher invited :)


Some tracing from logs [neutron, nova, ovn-sb-db]:

CONTROL-NODES:


ctl01-ovn-sb-db.log:2022-04-19T12:30:03.089Z|01002|raft|INFO|Transferring 
leadership to write a snapshot.
ctl01-ovn-sb-db.log:2022-04-19T12:30:03.099Z|01003|raft|INFO|server 1c5f is 
leader for term 42

ctl03-ovn-sb-db.log:2022-04-19T12:30:03.090Z|00938|raft|INFO|received 
leadership transfer from 1f46 in term 41
ctl03-ovn-sb-db.log:2022-04-19T12:30:03.092Z|00940|raft|INFO|term 42: elected 
leader by 2+ of 3 servers
ctl03-ovn-sb-db.log:2022-04-19T12:30:10.941Z|00941|jsonrpc|WARN|tcp:xx.yy.zz.26:41882:
 send error: Connection reset by peer
ctl03-ovn-sb-db.log:2022-04-19T12:30:27.324Z|00943|jsonrpc|WARN|tcp:xx.yy.zz.26:41896:
 send error: Connection reset by peer
ctl03-ovn-sb-db.log:2022-04-19T12:30:27.325Z|00945|jsonrpc|WARN|tcp:xx.yy.zz.26:41880:
 send error: Connection reset by peer
ctl03-ovn-sb-db.log:2022-04-19T12:30:27.325Z|00947|jsonrpc|WARN|tcp:xx.yy.zz.26:41892:
 send error: Connection reset by peer
ctl03-ovn-sb-db.log:2022-04-19T12:30:27.327Z|00949|jsonrpc|WARN|tcp:xx.yy.zz.26:41884:
 send error: Connection reset by peer
ctl03-ovn-sb-db.log:2022-04-19T12:31:49.244Z|00951|jsonrpc|WARN|tcp:xx.yy.zz.25:40260:
 send error: Connection timed out
ctl03-ovn-sb-db.log:2022-04-19T12:31:49.244Z|00953|jsonrpc|WARN|tcp:xx.yy.zz.25:40264:
 send error: Connection timed out
ctl03-ovn-sb-db.log:2022-04-19T12:31:49.244Z|00955|jsonrpc|WARN|tcp:xx.yy.zz.24:37440:
 send error: Connection timed out
ctl03-ovn-sb-db.log:2022-04-19T12:31:49.244Z|00957|jsonrpc|WARN|tcp:xx.yy.zz.24:37442:
 send error: Connection timed out
ctl03-ovn-sb-db.log:2022-04-19T12:31:49.245Z|00959|jsonrpc|WARN|tcp:xx.yy.zz.24:37446:
 send error: Connection timed out
ctl03-ovn-sb-db.log:2022-04-19T12:32:01.533Z|01001|jsonrpc|WARN|tcp:xx.yy.zz.67:57586:
 send error: Connection timed out


2022-04-19 12:30:08.898 27 INFO neutron.db.ovn_revision_numbers_db 
[req-7fcfdd74-482d-46b2-9f76-07190669d76d ff1516be452b4b939314bf3864a63f35 
9d3ae9a7b121488285203b0fdeabc3a3 - default default] Successfully bumped 
revision number for resource be178a9a-26d7-4bf0-a4e8-d206a6965205 (type: ports) 
to 1

2022-04-19 12:30:09.644 27 INFO neutron.db.ovn_revision_numbers_db
[req-a8278418-3ad9-450c-89bb-e7a5c1c0a06d
a9864cd890224c079051b3f56021be64 72db34087b9b401d842b66643b647e16 -
default default] Successfully bumped revision number for resource
be178a9a-26d7-4bf0-a4e8-d206a6965205 (type: ports) to 2


2022-04-19 12:30:10.235 27 INFO neutron.wsgi 
[req-571b53cc-ca04-46f7-89f9-fdf8e5931f4c a9864cd890224c079051b3f56021be64 
72db34087b9b401d842b66643b647e16 - default default] xx.yy.zz.68,xx.yy.zz.26 
"GET 
/v2.0/ports?tenant_id=9d3ae9a7b121488285203b0fdeabc3a3&device_id=7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34
 HTTP/1.1" status: 200  len: 1081 time: 0.0363808

2022-04-19 12:30:10.395 25 INFO
nova.api.openstack.compute.server_external_events
[req-d144458a-ca21-4e92-b5be-4f589e118d8c
4688fc44551b45c6a761c9c0351e76a8 72db34087b9b401d842b66643b647e16 -
default default] Creating event network-
changed:be178a9a-26d7-4bf0-a4e8-d206a6965205 for instance
7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34 on cmp-ucs-01-05.poc.example.com


2022-04-19 12:30:10.402 27 INFO neutron.notifiers.nova [-] Nova event response: 
{'name': 'network-changed', 'server_uuid': 
'7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34', 'tag': 
'be178a9a-26d7-4bf0-a4e8-d206a6965205', 'status': 'completed', 'code': 200}

2022-04-19 12:33:01.497 34 INFO neutron.db.ovn_revision_numbers_db
[req-82f3d4fe-1276-4918-86d3-55cc42ca2816 - - - - -] Successfully bumped
revision number for resource be178a9a-26d7-4bf0-a4e8-d206a6965205 (type:
ports) to 3

COMPUTE-NODES:

ovs-vswitchd.log:2022-04-19T12:30:10.743Z|08415|bridge|INFO|bridge br-int: 
added interface tapbe178a9a-26 on port 874
ovn-controller.log:2022-04-19T12:30:11.251Z|02356|binding|INFO|Claiming lport 
be178a9a-26d7-4bf0-a4e8-d206a6965205 for this chassis.
ovn-controller.log:2022-04-19T12:30:11.251Z|02357|binding|INFO|be178a9a-26d7-4bf0-a4e8-d206a6965205:
 Claiming fa:16:3e:8f:3b:47 1.80.1.254

nova-compute.log:2022-04-19 12:30:10.672 8 INFO os_vif
[req-2c8b5f82-ce4a-4aa7-9f0a-d346ee210ffb
ff1516be452b4b939314bf3864a63f35 9d3ae9a7b121488285203b0fdeabc3a3 -
default default] Successfully plugged vif
VIFOpenVSwitch(active=False,address=fa:16:3e:8f:3b:47,bridge_name='br-
int',has_traffic_filtering=True,id=be178a9a-26d7-4bf0-a4e8-d206a6965205,network=Network(5dd2512b-56f8-4e04-8623-971a633a76aa),plugin='ovs',port_profile=VIFPortProfileOpenVSwitch,preserve_on_delete=False,vif_name='tapbe178a9a-26')

ovs-vswitchd.log:2022-04-19T12:35:11.302Z|08448|bridge|INFO|bridge br-int: 
deleted interface tapbe178a9a-26 on port 874
ovn-controller.log:2022-04-19T12:35:11.307Z|02380|binding|INFO|Releasing lport 
be178a9a-26d7-4bf0-a4e8-d206a6965205 from this chassis.

2022-04-19 12:35:11.842 8 ERROR nova.compute.manager 
[req-2c8b5f82-ce4a-4aa7-9f0a-d346ee210ffb ff1516be452b4b939314bf3864a63f35 
9d3ae9a7b121488285203b0fdeabc3a3 - default default] [instance: 
7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34] Failed to
 allocate network(s): nova.exception.VirtualInterfaceCreateException: Virtual 
Interface creation failed
2022-04-19 12:35:11.842 8 ERROR nova.compute.manager [instance: 
7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34] Traceback (most recent call last):
2022-04-19 12:35:11.842 8 ERROR nova.compute.manager [instance: 
7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34]   File 
"/var/lib/kolla/venv/lib/python3.6/site-packages/nova/virt/libvirt/driver.py", 
line 7260, in _create_guest_with_network
2022-04-19 12:35:11.842 8 ERROR nova.compute.manager [instance: 
7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34]     post_xml_callback=post_xml_callback)
2022-04-19 12:35:11.842 8 ERROR nova.compute.manager [instance: 
7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34]   File 
"/usr/lib64/python3.6/contextlib.py", line 88, in __exit__
2022-04-19 12:35:11.842 8 ERROR nova.compute.manager [instance: 
7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34]     next(self.gen)
2022-04-19 12:35:11.842 8 ERROR nova.compute.manager [instance: 
7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34]   File 
"/var/lib/kolla/venv/lib/python3.6/site-packages/nova/compute/manager.py", line 
479, in wait_for_instance_event
2022-04-19 12:35:11.842 8 ERROR nova.compute.manager [instance: 
7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34]     actual_event = event.wait()
2022-04-19 12:35:11.842 8 ERROR nova.compute.manager [instance: 
7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34]   File 
"/var/lib/kolla/venv/lib/python3.6/site-packages/eventlet/event.py", line 125, 
in wait
2022-04-19 12:35:11.842 8 ERROR nova.compute.manager [instance: 
7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34]     result = hub.switch()
2022-04-19 12:35:11.842 8 ERROR nova.compute.manager [instance: 
7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34]   File 
"/var/lib/kolla/venv/lib/python3.6/site-packages/eventlet/hubs/hub.py", line 
313, in switch
2022-04-19 12:35:11.842 8 ERROR nova.compute.manager [instance: 
7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34]     return self.greenlet.switch()
2022-04-19 12:35:11.842 8 ERROR nova.compute.manager [instance: 
7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34] eventlet.timeout.Timeout: 300 seconds
2022-04-19 12:35:11.842 8 ERROR nova.compute.manager [instance: 
7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34] 
2022-04-19 12:35:11.842 8 ERROR nova.compute.manager [instance: 
7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34] During handling of the above exception, 
another exception occurred:
2022-04-19 12:35:11.842 8 ERROR nova.compute.manager [instance: 
7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34] 
2022-04-19 12:35:11.842 8 ERROR nova.compute.manager [instance: 
7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34] Traceback (most recent call last):
2022-04-19 12:35:11.842 8 ERROR nova.compute.manager [instance: 
7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34]   File 
"/var/lib/kolla/venv/lib/python3.6/site-packages/nova/compute/manager.py", line 
2402, in _build_and_run_instance
2022-04-19 12:35:11.842 8 ERROR nova.compute.manager [instance: 
7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34]     accel_info=accel_info)
2022-04-19 12:35:11.842 8 ERROR nova.compute.manager [instance: 
7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34]   File 
"/var/lib/kolla/venv/lib/python3.6/site-packages/nova/virt/libvirt/driver.py", 
line 4225, in spawn
2022-04-19 12:35:11.842 8 ERROR nova.compute.manager [instance: 
7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34]     cleanup_instance_disks=created_disks)
2022-04-19 12:35:11.842 8 ERROR nova.compute.manager [instance: 
7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34]   File 
"/var/lib/kolla/venv/lib/python3.6/site-packages/nova/virt/libvirt/driver.py", 
line 7283, in _create_guest_with_network
2022-04-19 12:35:11.842 8 ERROR nova.compute.manager [instance: 
7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34]     raise 
exception.VirtualInterfaceCreateException()
2022-04-19 12:35:11.842 8 ERROR nova.compute.manager [instance: 
7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34] 
nova.exception.VirtualInterfaceCreateException: Virtual Interface creation 
failed
2022-04-19 12:35:11.842 8 ERROR nova.compute.manager [instance: 
7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34] 
2022-04-19 12:35:11.844 8 ERROR nova.compute.manager 
[req-2c8b5f82-ce4a-4aa7-9f0a-d346ee210ffb ff1516be452b4b939314bf3864a63f35 
9d3ae9a7b121488285203b0fdeabc3a3 - default default] [instance: 
7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34] Build of 
instance 7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34 aborted: Failed to allocate the 
network(s), not rescheduling.: nova.exception.BuildAbortException: Build of 
instance 7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34 aborted: Failed to allocate the 
networ
k(s), not rescheduling.


CONTROL:

2022-04-19 12:35:11.330 25 INFO
neutron.plugins.ml2.drivers.ovn.mech_driver.mech_driver
[req-2b4d10b1-3a7a-4683-97df-990259705b46 - - - - -] OVN reports status
down for port: be178a9a-26d7-4bf0-a4e8-d206a6965205

2022-04-19 12:35:12.524 23 WARNING neutron.notifiers.nova [-] Nova
event: {'server_uuid': '7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34', 'name':
'network-vif-deleted', 'tag': 'be178a9a-26d7-4bf0-a4e8-d206a6965205',
'status': 'failed', 'code': 422} returned with failed status

/var/log/kolla/neutron/neutron-server.log:2022-04-19 12:35:13.016 25
WARNING neutron.notifiers.nova [-] Nova event: {'server_uuid':
'7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34', 'name': 'network-vif-plugged',
'status': 'failed', 'tag': 'be178a9a-26d7-4bf0-a4e8-d206a6965205',
'code': 422} returned with failed status

** Affects: neutron
     Importance: Undecided
         Status: New

** Affects: nova
     Importance: Undecided
         Status: New


** Tags: neutron ovn

** Also affects: nova
   Importance: Undecided
       Status: New

** Tags added: neutron ovn

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/1969592

Title:
  [OVN] Frequent DB leader changes causes 'VIF creation failed' on nova
  side

Status in neutron:
  New
Status in OpenStack Compute (nova):
  New

Bug description:
  Hello guys.

  As I found in 2021 there was a commit to OVS [since OVS v2.16 or 2.15 with 
that backport] that changed behavior during OVN DBs snapshoting. 
  Now before the leader creates a snapshot it will transfer leadership to 
another node. 

  We've run tests with rally and tempest and looks like there is a problem now 
when there is interaction between nova and neutron. 
  For example, simple rally test like 'create {network,router,subnet} -> add 
interface to router' looks okay even with 256 concurrent same tests/threads. 
But something like 'neutron.create_subnet -> nova.boot_server -> 
nova.attach_interface' will fail in time when transfer leadership happens. 
  Since it happens to often [ each 10m + rand(10) ] we will get a lot of 
errors. 

  This problem can be observed on all versions where OVS 2.16 [or
  backport] or higher invited :)

  
  Some tracing from logs [neutron, nova, ovn-sb-db]:

  CONTROL-NODES:

  
  ctl01-ovn-sb-db.log:2022-04-19T12:30:03.089Z|01002|raft|INFO|Transferring 
leadership to write a snapshot.
  ctl01-ovn-sb-db.log:2022-04-19T12:30:03.099Z|01003|raft|INFO|server 1c5f is 
leader for term 42

  ctl03-ovn-sb-db.log:2022-04-19T12:30:03.090Z|00938|raft|INFO|received 
leadership transfer from 1f46 in term 41
  ctl03-ovn-sb-db.log:2022-04-19T12:30:03.092Z|00940|raft|INFO|term 42: elected 
leader by 2+ of 3 servers
  
ctl03-ovn-sb-db.log:2022-04-19T12:30:10.941Z|00941|jsonrpc|WARN|tcp:xx.yy.zz.26:41882:
 send error: Connection reset by peer
  
ctl03-ovn-sb-db.log:2022-04-19T12:30:27.324Z|00943|jsonrpc|WARN|tcp:xx.yy.zz.26:41896:
 send error: Connection reset by peer
  
ctl03-ovn-sb-db.log:2022-04-19T12:30:27.325Z|00945|jsonrpc|WARN|tcp:xx.yy.zz.26:41880:
 send error: Connection reset by peer
  
ctl03-ovn-sb-db.log:2022-04-19T12:30:27.325Z|00947|jsonrpc|WARN|tcp:xx.yy.zz.26:41892:
 send error: Connection reset by peer
  
ctl03-ovn-sb-db.log:2022-04-19T12:30:27.327Z|00949|jsonrpc|WARN|tcp:xx.yy.zz.26:41884:
 send error: Connection reset by peer
  
ctl03-ovn-sb-db.log:2022-04-19T12:31:49.244Z|00951|jsonrpc|WARN|tcp:xx.yy.zz.25:40260:
 send error: Connection timed out
  
ctl03-ovn-sb-db.log:2022-04-19T12:31:49.244Z|00953|jsonrpc|WARN|tcp:xx.yy.zz.25:40264:
 send error: Connection timed out
  
ctl03-ovn-sb-db.log:2022-04-19T12:31:49.244Z|00955|jsonrpc|WARN|tcp:xx.yy.zz.24:37440:
 send error: Connection timed out
  
ctl03-ovn-sb-db.log:2022-04-19T12:31:49.244Z|00957|jsonrpc|WARN|tcp:xx.yy.zz.24:37442:
 send error: Connection timed out
  
ctl03-ovn-sb-db.log:2022-04-19T12:31:49.245Z|00959|jsonrpc|WARN|tcp:xx.yy.zz.24:37446:
 send error: Connection timed out
  
ctl03-ovn-sb-db.log:2022-04-19T12:32:01.533Z|01001|jsonrpc|WARN|tcp:xx.yy.zz.67:57586:
 send error: Connection timed out

  
  2022-04-19 12:30:08.898 27 INFO neutron.db.ovn_revision_numbers_db 
[req-7fcfdd74-482d-46b2-9f76-07190669d76d ff1516be452b4b939314bf3864a63f35 
9d3ae9a7b121488285203b0fdeabc3a3 - default default] Successfully bumped 
revision number for resource be178a9a-26d7-4bf0-a4e8-d206a6965205 (type: ports) 
to 1

  2022-04-19 12:30:09.644 27 INFO neutron.db.ovn_revision_numbers_db
  [req-a8278418-3ad9-450c-89bb-e7a5c1c0a06d
  a9864cd890224c079051b3f56021be64 72db34087b9b401d842b66643b647e16 -
  default default] Successfully bumped revision number for resource
  be178a9a-26d7-4bf0-a4e8-d206a6965205 (type: ports) to 2

  
  2022-04-19 12:30:10.235 27 INFO neutron.wsgi 
[req-571b53cc-ca04-46f7-89f9-fdf8e5931f4c a9864cd890224c079051b3f56021be64 
72db34087b9b401d842b66643b647e16 - default default] xx.yy.zz.68,xx.yy.zz.26 
"GET 
/v2.0/ports?tenant_id=9d3ae9a7b121488285203b0fdeabc3a3&device_id=7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34
 HTTP/1.1" status: 200  len: 1081 time: 0.0363808

  2022-04-19 12:30:10.395 25 INFO
  nova.api.openstack.compute.server_external_events
  [req-d144458a-ca21-4e92-b5be-4f589e118d8c
  4688fc44551b45c6a761c9c0351e76a8 72db34087b9b401d842b66643b647e16 -
  default default] Creating event network-
  changed:be178a9a-26d7-4bf0-a4e8-d206a6965205 for instance
  7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34 on cmp-ucs-01-05.poc.example.com

  
  2022-04-19 12:30:10.402 27 INFO neutron.notifiers.nova [-] Nova event 
response: {'name': 'network-changed', 'server_uuid': 
'7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34', 'tag': 
'be178a9a-26d7-4bf0-a4e8-d206a6965205', 'status': 'completed', 'code': 200}

  2022-04-19 12:33:01.497 34 INFO neutron.db.ovn_revision_numbers_db
  [req-82f3d4fe-1276-4918-86d3-55cc42ca2816 - - - - -] Successfully
  bumped revision number for resource
  be178a9a-26d7-4bf0-a4e8-d206a6965205 (type: ports) to 3

  COMPUTE-NODES:

  ovs-vswitchd.log:2022-04-19T12:30:10.743Z|08415|bridge|INFO|bridge br-int: 
added interface tapbe178a9a-26 on port 874
  ovn-controller.log:2022-04-19T12:30:11.251Z|02356|binding|INFO|Claiming lport 
be178a9a-26d7-4bf0-a4e8-d206a6965205 for this chassis.
  
ovn-controller.log:2022-04-19T12:30:11.251Z|02357|binding|INFO|be178a9a-26d7-4bf0-a4e8-d206a6965205:
 Claiming fa:16:3e:8f:3b:47 1.80.1.254

  nova-compute.log:2022-04-19 12:30:10.672 8 INFO os_vif
  [req-2c8b5f82-ce4a-4aa7-9f0a-d346ee210ffb
  ff1516be452b4b939314bf3864a63f35 9d3ae9a7b121488285203b0fdeabc3a3 -
  default default] Successfully plugged vif
  VIFOpenVSwitch(active=False,address=fa:16:3e:8f:3b:47,bridge_name='br-
  
int',has_traffic_filtering=True,id=be178a9a-26d7-4bf0-a4e8-d206a6965205,network=Network(5dd2512b-56f8-4e04-8623-971a633a76aa),plugin='ovs',port_profile=VIFPortProfileOpenVSwitch,preserve_on_delete=False,vif_name='tapbe178a9a-26')

  ovs-vswitchd.log:2022-04-19T12:35:11.302Z|08448|bridge|INFO|bridge br-int: 
deleted interface tapbe178a9a-26 on port 874
  ovn-controller.log:2022-04-19T12:35:11.307Z|02380|binding|INFO|Releasing 
lport be178a9a-26d7-4bf0-a4e8-d206a6965205 from this chassis.

  2022-04-19 12:35:11.842 8 ERROR nova.compute.manager 
[req-2c8b5f82-ce4a-4aa7-9f0a-d346ee210ffb ff1516be452b4b939314bf3864a63f35 
9d3ae9a7b121488285203b0fdeabc3a3 - default default] [instance: 
7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34] Failed to
   allocate network(s): nova.exception.VirtualInterfaceCreateException: Virtual 
Interface creation failed
  2022-04-19 12:35:11.842 8 ERROR nova.compute.manager [instance: 
7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34] Traceback (most recent call last):
  2022-04-19 12:35:11.842 8 ERROR nova.compute.manager [instance: 
7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34]   File 
"/var/lib/kolla/venv/lib/python3.6/site-packages/nova/virt/libvirt/driver.py", 
line 7260, in _create_guest_with_network
  2022-04-19 12:35:11.842 8 ERROR nova.compute.manager [instance: 
7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34]     post_xml_callback=post_xml_callback)
  2022-04-19 12:35:11.842 8 ERROR nova.compute.manager [instance: 
7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34]   File 
"/usr/lib64/python3.6/contextlib.py", line 88, in __exit__
  2022-04-19 12:35:11.842 8 ERROR nova.compute.manager [instance: 
7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34]     next(self.gen)
  2022-04-19 12:35:11.842 8 ERROR nova.compute.manager [instance: 
7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34]   File 
"/var/lib/kolla/venv/lib/python3.6/site-packages/nova/compute/manager.py", line 
479, in wait_for_instance_event
  2022-04-19 12:35:11.842 8 ERROR nova.compute.manager [instance: 
7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34]     actual_event = event.wait()
  2022-04-19 12:35:11.842 8 ERROR nova.compute.manager [instance: 
7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34]   File 
"/var/lib/kolla/venv/lib/python3.6/site-packages/eventlet/event.py", line 125, 
in wait
  2022-04-19 12:35:11.842 8 ERROR nova.compute.manager [instance: 
7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34]     result = hub.switch()
  2022-04-19 12:35:11.842 8 ERROR nova.compute.manager [instance: 
7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34]   File 
"/var/lib/kolla/venv/lib/python3.6/site-packages/eventlet/hubs/hub.py", line 
313, in switch
  2022-04-19 12:35:11.842 8 ERROR nova.compute.manager [instance: 
7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34]     return self.greenlet.switch()
  2022-04-19 12:35:11.842 8 ERROR nova.compute.manager [instance: 
7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34] eventlet.timeout.Timeout: 300 seconds
  2022-04-19 12:35:11.842 8 ERROR nova.compute.manager [instance: 
7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34] 
  2022-04-19 12:35:11.842 8 ERROR nova.compute.manager [instance: 
7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34] During handling of the above exception, 
another exception occurred:
  2022-04-19 12:35:11.842 8 ERROR nova.compute.manager [instance: 
7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34] 
  2022-04-19 12:35:11.842 8 ERROR nova.compute.manager [instance: 
7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34] Traceback (most recent call last):
  2022-04-19 12:35:11.842 8 ERROR nova.compute.manager [instance: 
7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34]   File 
"/var/lib/kolla/venv/lib/python3.6/site-packages/nova/compute/manager.py", line 
2402, in _build_and_run_instance
  2022-04-19 12:35:11.842 8 ERROR nova.compute.manager [instance: 
7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34]     accel_info=accel_info)
  2022-04-19 12:35:11.842 8 ERROR nova.compute.manager [instance: 
7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34]   File 
"/var/lib/kolla/venv/lib/python3.6/site-packages/nova/virt/libvirt/driver.py", 
line 4225, in spawn
  2022-04-19 12:35:11.842 8 ERROR nova.compute.manager [instance: 
7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34]     cleanup_instance_disks=created_disks)
  2022-04-19 12:35:11.842 8 ERROR nova.compute.manager [instance: 
7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34]   File 
"/var/lib/kolla/venv/lib/python3.6/site-packages/nova/virt/libvirt/driver.py", 
line 7283, in _create_guest_with_network
  2022-04-19 12:35:11.842 8 ERROR nova.compute.manager [instance: 
7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34]     raise 
exception.VirtualInterfaceCreateException()
  2022-04-19 12:35:11.842 8 ERROR nova.compute.manager [instance: 
7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34] 
nova.exception.VirtualInterfaceCreateException: Virtual Interface creation 
failed
  2022-04-19 12:35:11.842 8 ERROR nova.compute.manager [instance: 
7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34] 
  2022-04-19 12:35:11.844 8 ERROR nova.compute.manager 
[req-2c8b5f82-ce4a-4aa7-9f0a-d346ee210ffb ff1516be452b4b939314bf3864a63f35 
9d3ae9a7b121488285203b0fdeabc3a3 - default default] [instance: 
7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34] Build of 
  instance 7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34 aborted: Failed to allocate the 
network(s), not rescheduling.: nova.exception.BuildAbortException: Build of 
instance 7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34 aborted: Failed to allocate the 
networ
  k(s), not rescheduling.

  
  CONTROL:

  2022-04-19 12:35:11.330 25 INFO
  neutron.plugins.ml2.drivers.ovn.mech_driver.mech_driver
  [req-2b4d10b1-3a7a-4683-97df-990259705b46 - - - - -] OVN reports
  status down for port: be178a9a-26d7-4bf0-a4e8-d206a6965205

  2022-04-19 12:35:12.524 23 WARNING neutron.notifiers.nova [-] Nova
  event: {'server_uuid': '7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34', 'name':
  'network-vif-deleted', 'tag': 'be178a9a-26d7-4bf0-a4e8-d206a6965205',
  'status': 'failed', 'code': 422} returned with failed status

  /var/log/kolla/neutron/neutron-server.log:2022-04-19 12:35:13.016 25
  WARNING neutron.notifiers.nova [-] Nova event: {'server_uuid':
  '7560fbb7-3ec7-41ef-b7a5-5e955ca4ff34', 'name': 'network-vif-plugged',
  'status': 'failed', 'tag': 'be178a9a-26d7-4bf0-a4e8-d206a6965205',
  'code': 422} returned with failed status

To manage notifications about this bug go to:
https://bugs.launchpad.net/neutron/+bug/1969592/+subscriptions


-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to     : [email protected]
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp

Reply via email to