[Yahoo-eng-team] [Bug 2035573] [NEW] Insertion of a duplicated ProviderResourceAssociation entry while creating a HA router

2023-09-13 Thread Jimin Shin
Public bug reported:

[SUMMARY]
While creating multiple HA routers in a row, some router creations are failed 
with the log below.

2023-07-21 02:28:19.968 12 DEBUG neutron.wsgi [-] (12) accepted ('10.2.41.158', 
57438) server 
/var/lib/kolla/venv/lib/python3.10/site-packages/eventlet/wsgi.py:1004
2023-07-21 02:28:19.979 12 DEBUG neutron.api.v2.base [None 
req-726ab30e-b9c9-4e38-8c9d-216752ab2aea 439ae1ccaa9a494284cee3fdb6227208 
97895007888245c3acdfc41146d2e151 - - default default] Request body: {'router': 
{'name': 'test6', 'admin_state_up': True, 'tenant_id': 
'97895007888245c3acdfc41146d2e151'}} prepare_request_body 
/var/lib/kolla/venv/lib/python3.10/site-packages/neutron/api/v2/base.py:731
2023-07-21 02:28:20.178 12 ERROR neutron.db.l3_hamode_db [None 
req-726ab30e-b9c9-4e38-8c9d-216752ab2aea 439ae1ccaa9a494284cee3fdb6227208 
97895007888245c3acdfc41146d2e151 - - default default] Failed to schedule HA 
router 5a599ade-7b6a-4b3e-b635-ca00e37f2657.: 
neutron_lib.objects.exceptions.NeutronDbObjectDuplicateEntry: Failed to create 
a duplicate ProviderResourceAssociation: for attribute(s) ['PRIMARY'] with 
value(s) ha-5a599ade-7b6a-4b3e-b635-ca00e37f2657
2023-07-21 02:28:20.178 12 ERROR neutron.db.l3_hamode_db Traceback (most recent 
call last):
2023-07-21 02:28:20.178 12 ERROR neutron.db.l3_hamode_db   File 
"/var/lib/kolla/venv/lib/python3.10/site-packages/sqlalchemy/engine/base.py", 
line 1900, in _execute_context
2023-07-21 02:28:20.178 12 ERROR neutron.db.l3_hamode_db     
self.dialect.do_execute(
2023-07-21 02:28:20.178 12 ERROR neutron.db.l3_hamode_db   File 
"/var/lib/kolla/venv/lib/python3.10/site-packages/sqlalchemy/engine/default.py",
 line 736, in do_execute
2023-07-21 02:28:20.178 12 ERROR neutron.db.l3_hamode_db     
cursor.execute(statement, parameters)
2023-07-21 02:28:20.178 12 ERROR neutron.db.l3_hamode_db   File 
"/var/lib/kolla/venv/lib/python3.10/site-packages/pymysql/cursors.py", line 
148, in execute
2023-07-21 02:28:20.178 12 ERROR neutron.db.l3_hamode_db     result = 
self._query(query)
2023-07-21 02:28:20.178 12 ERROR neutron.db.l3_hamode_db   File 
"/var/lib/kolla/venv/lib/python3.10/site-packages/pymysql/cursors.py", line 
310, in _query
2023-07-21 02:28:20.178 12 ERROR neutron.db.l3_hamode_db     conn.query(q)
2023-07-21 02:28:20.178 12 ERROR neutron.db.l3_hamode_db   File 
"/var/lib/kolla/venv/lib/python3.10/site-packages/pymysql/connections.py", line 
548, in query
2023-07-21 02:28:20.178 12 ERROR neutron.db.l3_hamode_db     
self._affected_rows = self._read_query_result(unbuffered=unbuffered)
2023-07-21 02:28:20.178 12 ERROR neutron.db.l3_hamode_db   File 
"/var/lib/kolla/venv/lib/python3.10/site-packages/pymysql/connections.py", line 
775, in _read_query_result
2023-07-21 02:28:20.178 12 ERROR neutron.db.l3_hamode_db     result.read()
2023-07-21 02:28:20.178 12 ERROR neutron.db.l3_hamode_db   File 
"/var/lib/kolla/venv/lib/python3.10/site-packages/pymysql/connections.py", line 
1156, in read
2023-07-21 02:28:20.178 12 ERROR neutron.db.l3_hamode_db     first_packet = 
self.connection._read_packet()
2023-07-21 02:28:20.178 12 ERROR neutron.db.l3_hamode_db   File 
"/var/lib/kolla/venv/lib/python3.10/site-packages/pymysql/connections.py", line 
725, in _read_packet
2023-07-21 02:28:20.178 12 ERROR neutron.db.l3_hamode_db     
packet.raise_for_error()
2023-07-21 02:28:20.178 12 ERROR neutron.db.l3_hamode_db   File 
"/var/lib/kolla/venv/lib/python3.10/site-packages/pymysql/protocol.py", line 
221, in raise_for_error
2023-07-21 02:28:20.178 12 ERROR neutron.db.l3_hamode_db     
err.raise_mysql_exception(self._data)
2023-07-21 02:28:20.178 12 ERROR neutron.db.l3_hamode_db   File 
"/var/lib/kolla/venv/lib/python3.10/site-packages/pymysql/err.py", line 143, in 
raise_mysql_exception
2023-07-21 02:28:20.178 12 ERROR neutron.db.l3_hamode_db     raise 
errorclass(errno, errval)
2023-07-21 02:28:20.178 12 ERROR neutron.db.l3_hamode_db 
pymysql.err.IntegrityError: (1062, "Duplicate entry 
'ha-5a599ade-7b6a-4b3e-b635-ca00e37f2657' for key 'PRIMARY'")
2023-07-21 02:28:20.178 12 ERROR neutron.db.l3_hamode_db
2023-07-21 02:28:20.178 12 ERROR neutron.db.l3_hamode_db The above exception 
was the direct cause of the following exception:
2023-07-21 02:28:20.178 12 ERROR neutron.db.l3_hamode_db
2023-07-21 02:28:20.178 12 ERROR neutron.db.l3_hamode_db Traceback (most recent 
call last):
2023-07-21 02:28:20.178 12 ERROR neutron.db.l3_hamode_db   File 
"/var/lib/kolla/venv/lib/python3.10/site-packages/neutron/objects/base.py", 
line 903, in create
2023-07-21 02:28:20.178 12 ERROR neutron.db.l3_hamode_db     db_obj = 
obj_db_api.create_object(
2023-07-21 02:28:20.178 12 ERROR neutron.db.l3_hamode_db   File 
"/var/lib/kolla/venv/lib/python3.10/site-packages/neutron/objects/db/api.py", 
line 72, in create_object
2023-07-21 02:28:20.178 12 ERROR neutron.db.l3_hamode_db     with 
obj_cls.db_context_writer(context):
2023-07-21 02:28:20.178 12 ERROR neutron.db.l3_hamode_db   File 

[Yahoo-eng-team] [Bug 1943639] Re: project/instances/attach_interface has O(N) scaling time complexity for opening form

2023-09-13 Thread Corey Bryant
** Changed in: horizon (Ubuntu)
   Status: New => Fix Released

** Changed in: cloud-archive
   Status: New => Fix Released

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Dashboard (Horizon).
https://bugs.launchpad.net/bugs/1943639

Title:
  project/instances/attach_interface has O(N) scaling time complexity
  for opening form

Status in Ubuntu Cloud Archive:
  Fix Released
Status in Ubuntu Cloud Archive ussuri series:
  New
Status in Ubuntu Cloud Archive victoria series:
  New
Status in Ubuntu Cloud Archive wallaby series:
  New
Status in Ubuntu Cloud Archive xena series:
  Fix Released
Status in OpenStack Dashboard (Horizon):
  Fix Released
Status in horizon package in Ubuntu:
  Fix Released
Status in horizon source package in Focal:
  New

Bug description:
  [ Impact ]

  The time complexity of opening the project/instances/attach_interface
  form box is O(N) where N is the number of networks in the project with
  a large prefactor.

  This is due to

  
https://opendev.org/openstack/horizon/src/branch/master/openstack_dashboard/dashboards/project/instances/utils.py#L210

  Which loops over the networks and requests the ports associated with
  the network. For large projects this scaling behavior can become
  prohibitive.

  The patch [1] addresses this issue by reducing the number of API calls
  and hence the prefactor of the algorithm.

  [ Test Plan ]

  In order to reproduce the issue, create a Nova VM and then add many
  networks. On the instances tab in the Horizon UI click on "attach
  interface" for the VM. It will take a moment for the dialog to appear.
  The exact time until the dialog appears will depend on the number of
  networks linearly.

  With [1] the time it takes for the dialog box to appear will be
  significantly shorter.

  [ Where problems could occur ]

  The patch [1] affects the "attach interface" dialog box and could
  break this UI feature in case something was wrong with the
  implementation. It is also possible that due to a bug in the
  implementation some networks are missing from the dialog.

  [1] https://review.opendev.org/c/openstack/horizon/+/866895

To manage notifications about this bug go to:
https://bugs.launchpad.net/cloud-archive/+bug/1943639/+subscriptions


-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 2035382] [NEW] [FT] "TestMonitorDaemon._run_monitor" failing radomly, initial message not written

2023-09-13 Thread Rodolfo Alonso
Public bug reported:

The method ``TestMonitorDaemon._run_monitor`` is randomly failing. The
initial message expected after "keepalived_state_change" process is
started is not written in the logs.

Logs:
https://8271f43479f81f7d3395-4bdeef087e9d95514555f2932706a956.ssl.cf1.rackcdn.com/893555/1/check/neutron-
functional-with-uwsgi/90dde54/testr_results.html

Snippet: https://paste.opendev.org/show/bJCAjIqiUOcWB0xBhPnX/

** Affects: neutron
 Importance: Undecided
 Status: New

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/2035382

Title:
  [FT] "TestMonitorDaemon._run_monitor" failing radomly, initial message
  not written

Status in neutron:
  New

Bug description:
  The method ``TestMonitorDaemon._run_monitor`` is randomly failing. The
  initial message expected after "keepalived_state_change" process is
  started is not written in the logs.

  Logs:
  
https://8271f43479f81f7d3395-4bdeef087e9d95514555f2932706a956.ssl.cf1.rackcdn.com/893555/1/check/neutron-
  functional-with-uwsgi/90dde54/testr_results.html

  Snippet: https://paste.opendev.org/show/bJCAjIqiUOcWB0xBhPnX/

To manage notifications about this bug go to:
https://bugs.launchpad.net/neutron/+bug/2035382/+subscriptions


-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 2035375] [NEW] Detaching multiple NVMe-oF volumes may leave the subsystem in connecting state

2023-09-13 Thread Gorka Eguileor
Public bug reported:

When detaching multiple NVMe-oF volumes from the same host we may end
with a NVMe subsystem in "connecting" state, and we'll see a bunch nvme
error in dmesg.

This happens on storage systems that share the same subsystem for
multiple volumes because Nova has not been updated to support the tri-
state "shared_targets" option that groups the detach and unmap of
volumes to prevent race conditions.

This is related to the issue mentioned in an os-brick commit message:
https://review.opendev.org/c/openstack/os-brick/+/836062/12//COMMIT_MSG

** Affects: nova
 Importance: Undecided
 Assignee: Gorka Eguileor (gorka)
 Status: New

** Changed in: nova
 Assignee: (unassigned) => Gorka Eguileor (gorka)

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/2035375

Title:
  Detaching multiple NVMe-oF volumes may leave the subsystem in
  connecting state

Status in OpenStack Compute (nova):
  New

Bug description:
  When detaching multiple NVMe-oF volumes from the same host we may end
  with a NVMe subsystem in "connecting" state, and we'll see a bunch
  nvme error in dmesg.

  This happens on storage systems that share the same subsystem for
  multiple volumes because Nova has not been updated to support the tri-
  state "shared_targets" option that groups the detach and unmap of
  volumes to prevent race conditions.

  This is related to the issue mentioned in an os-brick commit message:
  https://review.opendev.org/c/openstack/os-
  brick/+/836062/12//COMMIT_MSG

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/2035375/+subscriptions


-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 2035368] [NEW] Debug config option in cinder and glance sections not working

2023-09-13 Thread Gorka Eguileor
Public bug reported:

Nova has the hability of individually setting debug mode for cinder
related libraries (cinderclient and os-brick) and for glanceclient using
their respective configuration sections and setting `debug = true`
regardless of the default debug setting.

Unfortunately these options don't work as expected and have no effect.

** Affects: nova
 Importance: Undecided
 Assignee: Gorka Eguileor (gorka)
 Status: New

** Changed in: nova
 Assignee: (unassigned) => Gorka Eguileor (gorka)

** Description changed:

- Nova has the possibility of individually setting debug mode for cinder
+ Nova has the hability of individually setting debug mode for cinder
  related libraries (cinderclient and os-brick) and for glanceclient using
  their respective configuration sections and setting `debug = true`
  regardless of the default debug setting.
  
  Unfortunately these options don't work as expected and have no effect.

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/2035368

Title:
  Debug config option in cinder and glance sections not working

Status in OpenStack Compute (nova):
  New

Bug description:
  Nova has the hability of individually setting debug mode for cinder
  related libraries (cinderclient and os-brick) and for glanceclient
  using their respective configuration sections and setting `debug =
  true` regardless of the default debug setting.

  Unfortunately these options don't work as expected and have no effect.

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/2035368/+subscriptions


-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1999209] Re: Internal ip of FIP will not changed if updated fixed_ips of internal port

2023-09-13 Thread OpenStack Infra
Reviewed:  https://review.opendev.org/c/openstack/neutron/+/889791
Committed: 
https://opendev.org/openstack/neutron/commit/aad82233eb4fa7a6354f58cd9f8fda869f77db80
Submitter: "Zuul (22348)"
Branch:master

commit aad82233eb4fa7a6354f58cd9f8fda869f77db80
Author: liushy 
Date:   Thu Jul 27 12:48:34 2023 +0800

Prevent internal IP change for floating IP

Raise an error when deleting/changing the fixed IP
which is linked to a floating IP.

Closes-Bug: #1999209
Change-Id: I83a5b6c30d54435426f75f4cd1f80bf41822eec5


** Changed in: neutron
   Status: In Progress => Fix Released

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/1999209

Title:
  Internal ip of FIP will not changed if updated fixed_ips of internal
  port

Status in neutron:
  Fix Released

Bug description:
  As the title describes, if we updated fixed_ips of one internal port
  which associated a floatingip, but the dnat_adn_snat entry in ovn will
  not changed.

To manage notifications about this bug go to:
https://bugs.launchpad.net/neutron/+bug/1999209/+subscriptions


-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 2034522] Re: Fake members operating_status ONLINE

2023-09-13 Thread OpenStack Infra
Reviewed:  https://review.opendev.org/c/openstack/ovn-octavia-provider/+/893839
Committed: 
https://opendev.org/openstack/ovn-octavia-provider/commit/fe6612f71475413b64e554b9a6707e35e9e5c053
Submitter: "Zuul (22348)"
Branch:master

commit fe6612f71475413b64e554b9a6707e35e9e5c053
Author: Fernando Royo 
Date:   Wed Sep 6 09:50:19 2023 +0200

Cover the use case of a member non existing

When a HM is attached to a pool and a backend member in that pool
is a fake member (e.g. due to a typo on creation) the member remains
in ONLINE status. Basically this is due to the fact that there
isn't any LSP attached to that member and no Service_Monitor entries
will take care of it.

This patch checks inmediatelly after creation the member and update
the whole LB status to reflect this fake member that could help to
the user to identify quickly those fake members.

Closes-Bug: 2034522
Change-Id: I72b2d9c5f454f9b156414bf91ca7deb7f0e9d8b0


** Changed in: neutron
   Status: In Progress => Fix Released

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/2034522

Title:
  Fake members operating_status ONLINE

Status in neutron:
  Fix Released

Bug description:
  I can deploy members with an invalid / invented ip address (no real
  servers with that address) and the LB shows that everything is ok with
  them (running `openstack loadbalancer status show ` will show that
  the members have "operating_status": "ONLINE".

  An example: I deployed the following:
  {
  "loadbalancer": {
  "id": "c50b7cb3-6b8f-434b-9a47-a10a27d0a9b5",
  "name": "ovn_lb",
  "operating_status": "ONLINE",
  "provisioning_status": "ACTIVE",
  "listeners": [
  {
  "id": "87bafdda-0ac6-438f-8824-cb75f9e014df",
  "name": "tcp_listener",
  "operating_status": "ONLINE",
  "provisioning_status": "ACTIVE",
  "pools": [
  {
  "id": "aa6ed64c-4d19-448b-969d-6cc686385162",
  "name": "tcp_pool",
  "provisioning_status": "ACTIVE",
  "operating_status": "ONLINE",
  "health_monitor": {
  "id": "cc72e7eb-722b-49be-b3d2-3857f880346d",
  "name": "hm_ovn_provider",
  "type": "TCP",
  "provisioning_status": "ACTIVE",
  "operating_status": "ONLINE"
  },
  "members": [
  {
  "id": "648b9d51-115a-4312-b92e-cc59af0d0401",
  "name": "fake_member",
  "operating_status": "ONLINE",
  "provisioning_status": "ACTIVE",
  "address": "10.100.0.204",
  "protocol_port": 80
  },
  {
  "id": "8dae11a2-e2d5-45f9-9e85-50f61fa07753",
  "name": "de3f2a06",
  "operating_status": "ONLINE",
  "provisioning_status": "ACTIVE",
  "address": "10.0.64.34",
  "protocol_port": 80
  },
  {
  "id": "9b044180-71b4-4fa6-83df-4d0f99b4a3f7",
  "name": "fake_member2",
  "operating_status": "ONLINE",
  "provisioning_status": "ACTIVE",
  "address": "10.100.0.205",
  "protocol_port": 80
  },
  {
  "id": "fe9ce8ca-e6b7-4c5b-807c-8e295156df85",
  "name": "6c186a80",
  "operating_status": "ONLINE",
  "provisioning_status": "ACTIVE",
  "address": "10.0.64.39",
  "protocol_port": 80
  }
  ]
  }
  ]
  }
  ]
  }
  }

  when the existing servers are the following:
  
+--+-++--+++
  | ID   | Name| Status | Networks  
 

[Yahoo-eng-team] [Bug 2035332] [NEW] VLAN networks for North / South Traffic Broken

2023-09-13 Thread Graeme Moss
Public bug reported:

## Environment

### Deployment

- Ubuntu 22.04 LTS
- Openstack Release ZED
- Kolla-ansible - stable/zed repo
- Kolla - stable/zed repo
- Containers built with ubuntu 22.04 LTS
- Containers built on 2023-08-23
- OVN+DVR+VLAN tenant networks.
- We have three controllers occ1, occ2 occ3
- Neutron version neutron-21.1.3.dev34 commit 
d6ee668cc32725cb7d15d2e08fdb50a761f91fe4
- ovn-nbctl 22.09.1
- Open vSwitch Library 3.0.3
- DB Schema 6.3.0

1.  New provider network deployed into openstack, on vlan 504.
2.  Router connected to this provider network.
3.  Instance connected to provider network no FIP

## Issues

Attempting to send north/south traffic (ping 8.8.8.8), results in the
following symptoms. 2 pings are successful, all other pings fail, until
the ping is cancelled, and a couple of minutes pass, then two pings will
be successful again, then back to failing.

New routers with vlan networks attached don't create all three ports on
the controllers.

Even when fixing the localnet ports on the router to have three with
changing the priority when attaching a FIP the N/S traffic is limited to
2 pings

Only when setting `reside-on-redirect-chassis` to `True` can we get the
vlan to work with FIP and have baremetal nodes have FIP.

## Diagnostics

After looking at the ovn-controller logs on the control nodes we can see
that it tries to claim the port on occ0001. which matches the gateway
chassis on the routers LRP port.

```
2023-09-06T14:13:32.454Z|00718|binding|INFO|Claiming lport 
cr-lrp-1a089d8f-d7a3-4116-a496-94cb87abe57f for this chassis.
2023-09-06T14:13:32.454Z|00719|binding|INFO|cr-lrp-1a089d8f-d7a3-4116-a496-94cb87abe57f:
 Claiming fa:16:3e:fc:ba:cf 1xx.xx.xxx.xxx/25
```

Gateway chassis of the LRP port.

```
ovn-nbctl list Gateway_Chassis | grep -A2 -B4 
lrp-71cf7286-de37-4d86-b362-eb7ba689d2d1

_uuid   : cf26be06-206d-443c-b224-25cc06ef2094
chassis_name: occ2
external_ids: {}
name: lrp-71cf7286-de37-4d86-b362-eb7ba689d2d1_occ2
options : {}
priority: 2
--

_uuid   : 1d9e8314-ed00-4694-8974-0328b78d34e1
chassis_name: occ1
external_ids: {}
name: lrp-71cf7286-de37-4d86-b362-eb7ba689d2d1_occ1
options : {}
priority: 3
--

_uuid   : b1e41ceb-ca2d-42eb-a896-b3551ea1b32f
chassis_name: occ3
external_ids: {}
name: lrp-71cf7286-de37-4d86-b362-eb7ba689d2d1_occ3
options : {}
priority: 1
```

We see nothing about `occ2` or `occ3` trying to claim the LRP port but 
we found that when you change the priority around to try resolve, we can see 
that the port is not on `occ1` but is on occ0002
We change occ0001 = 1 and occ0003 = 3 which means `occ3` will be come the 
highest gateway.

```
ovn-nbctl set gateway_chassis 1d9e8314-ed00-4694-8974-0328b78d34e1 priority=1
ovn-nbctl set gateway_chassis b1e41ceb-ca2d-42eb-a896-b3551ea1b32f priority=3
```

the logs show the following.

occ0001

```
2023-09-06T14:10:06.134Z|00667|binding|INFO|Releasing lport 
cr-lrp-71cf7286-de37-4d86-b362-eb7ba689d2d1 from this chassis (sb_readonly=0)
2023-09-06T14:10:06.134Z|00668|if_status|WARN|Trying to release unknown 
interface cr-lrp-71cf7286-de37-4d86-b362-eb7ba689d2d1
```

occ0002

```
2023-09-06T14:10:14.883Z|00444|binding|INFO|Releasing lport 
cr-lrp-71cf7286-de37-4d86-b362-eb7ba689d2d1 from this chassis (sb_readonly=0)
2023-09-06T14:10:14.883Z|00445|if_status|WARN|Trying to release unknown 
interface cr-lrp-71cf7286-de37-4d86-b362-eb7ba689d2d1
```

occ0003

```
2023-09-06T14:10:14.789Z|00459|binding|INFO|Changing chassis for lport 
cr-lrp-71cf7286-de37-4d86-b362-eb7ba689d2d1 from occ2 to occ3.
2023-09-06T14:10:14.789Z|00460|binding|INFO|cr-lrp-71cf7286-de37-4d86-b362-eb7ba689d2d1:
 Claiming fa:16:3e:71:df:71 1xx.xx.xxx.xxx/25
```

on `occ3` we can see that `occ2` had the gateway and not
`occ1` which it should of had. This happens on creating new routers
on the vlan provider network.All exisiting Routers before upgrade are
working and that they have the same priority.

## Second diagnostics

Looking at each Logical Router we can see that when the router is first created 
that only two of the three ports are created.
Broken router:

```
_uuid   : 773bb527-f193-4b47-8685-e62c9325dd7b
copp: []
enabled : true
external_ids: {"neutron:availability_zone_hints"="", 
"neutron:gw_network_id"="c9d130bc-301d-45c0-9328-a6964af65579", 
"neutron:gw_port_id"="1a089d8f-d7a3-4116-a496-94cb87abe57f", 
"neutron:revision_number"="4", "neutron:router_name"=new-r1-test}
load_balancer   : []
load_balancer_group : []
name: neutron-2b51e12e-5505-477e-9720-e5db31a05790
nat : [f22e6004-ad69-4b12-9445-7006a03495f5]
options : {always_learn_from_arp_request="false", 

[Yahoo-eng-team] [Bug 2035325] [NEW] FDB entries grows indefinitely

2023-09-13 Thread Luis Tomas
Public bug reported:

With the added support for learning FDB entries [1] there is a problem
that FDB table can grow indefinitely, leading to performance/scale
issues. New options are added to OVN [2] to tackle this problem, and
neutron should make use of it


[1] https://review.opendev.org/c/openstack/neutron/+/877675  
[2] 
https://github.com/ovn-org/ovn/commit/ae9a5488824c49e25215b02e7e81a62eb4d0bd53

** Affects: neutron
 Importance: Undecided
 Assignee: Luis Tomas (luis5tb)
 Status: In Progress

** Changed in: neutron
 Assignee: (unassigned) => Luis Tomas (luis5tb)

** Changed in: neutron
   Status: New => In Progress

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/2035325

Title:
  FDB entries grows indefinitely

Status in neutron:
  In Progress

Bug description:
  With the added support for learning FDB entries [1] there is a problem
  that FDB table can grow indefinitely, leading to performance/scale
  issues. New options are added to OVN [2] to tackle this problem, and
  neutron should make use of it

  
  [1] https://review.opendev.org/c/openstack/neutron/+/877675  
  [2] 
https://github.com/ovn-org/ovn/commit/ae9a5488824c49e25215b02e7e81a62eb4d0bd53

To manage notifications about this bug go to:
https://bugs.launchpad.net/neutron/+bug/2035325/+subscriptions


-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1983863] Re: Can't log within tpool.execute

2023-09-13 Thread Sylvain Bauza
** Changed in: nova
   Status: In Progress => Invalid

** Changed in: nova
   Status: Invalid => Fix Committed

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1983863

Title:
  Can't log within tpool.execute

Status in OpenStack Compute (nova):
  Fix Committed
Status in oslo.log:
  Fix Released

Bug description:
  There is a bug in eventlet where logging within a native thread can
  lead to a deadlock situation:
  https://github.com/eventlet/eventlet/issues/432

  When encountered with this issue some projects in OpenStack using
  oslo.log, eg. Cinder, resolve them by removing any logging withing
  native threads.

  There is actually a better approach.  The Swift team came up with a
  solution a long time ago, and it would be great if oslo.log could use
  this workaround automaticaly:
  
https://opendev.org/openstack/swift/commit/69c715c505cf9e5df29dc1dff2fa1a4847471cb6

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1983863/+subscriptions


-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp