[Kernel-packages] [Bug 2063315] Re: Suspend & Resume functionality broken/timesout in GCE

2024-04-24 Thread Philip Roche
** Also affects: linux-gcp (Ubuntu Noble)
   Importance: Undecided
   Status: New

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux-gcp in Ubuntu.
https://bugs.launchpad.net/bugs/2063315

Title:
  Suspend & Resume functionality broken/timesout in GCE

Status in Release Notes for Ubuntu:
  New
Status in linux-gcp package in Ubuntu:
  New
Status in linux-gcp source package in Noble:
  New

Bug description:
  [Impact]
   
  Suspend/Resume capability is broken in all noble images with kernel version 
6.8.0-1007-gcp.

  GCE offers the capability to "Suspend" a VM to conserve power/lower
  costs when the instance is not in use [0]. It uses ACPI S3 signals to
  tell the guest to power down. This capability no longer works in the
  latest kernel with the following error:

  ```
  Operation type [suspend] failed with message "Instance suspend failed due to 
guest timeout."
  ```

  which points to the following [1].

  

  Refs:

  [0]: https://cloud.google.com/compute/docs/instances/suspend-resume-
  instance

  [1]:
  https://cloud.google.com/compute/docs/troubleshooting/troubleshooting-
  suspend-resume#there_was_a_guest_timeout

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu-release-notes/+bug/2063315/+subscriptions


-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 2061851] Re: linux-gcp 6.8.0-1005.5 (+ others) Noble kernel regression with new apparmor profiles/features

2024-04-22 Thread Philip Roche
** Changed in: snapd (Ubuntu Noble)
   Status: New => Invalid

** Changed in: linux-aws (Ubuntu Noble)
   Status: New => Fix Released

** Changed in: linux-azure (Ubuntu Noble)
   Status: New => Fix Released

** Changed in: linux-gcp (Ubuntu Noble)
   Status: New => Fix Released

** Changed in: linux-ibm (Ubuntu Noble)
   Status: New => Fix Released

** Changed in: linux-oracle (Ubuntu Noble)
   Status: New => Fix Released

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/2061851

Title:
  linux-gcp 6.8.0-1005.5 (+ others) Noble kernel regression with new
  apparmor profiles/features

Status in chrony package in Ubuntu:
  Invalid
Status in linux package in Ubuntu:
  Fix Released
Status in linux-aws package in Ubuntu:
  Fix Released
Status in linux-azure package in Ubuntu:
  Fix Released
Status in linux-gcp package in Ubuntu:
  Fix Released
Status in linux-ibm package in Ubuntu:
  Fix Released
Status in linux-oracle package in Ubuntu:
  Fix Released
Status in snapd package in Ubuntu:
  Invalid
Status in chrony source package in Noble:
  Invalid
Status in linux source package in Noble:
  Fix Released
Status in linux-aws source package in Noble:
  Fix Released
Status in linux-azure source package in Noble:
  Fix Released
Status in linux-gcp source package in Noble:
  Fix Released
Status in linux-ibm source package in Noble:
  Fix Released
Status in linux-oracle source package in Noble:
  Fix Released
Status in snapd source package in Noble:
  Invalid

Bug description:
  * Canonical Public Cloud discovered that `chronyc -c sources` now fails with 
`506 Cannot talk to daemon` with the latest kernels. We are seeing this in 
linux-azure and linux-gcp kernels (6.8.0-1005.5)
  * Disabling AppArmor (`sudo systemctl stop apparmor`) completely results in 
no regression and `chronyc -c sources` returns as expected
  * Disabling the apparmor profile for `chronyd` only results in no regression 
and `chronyc -c sources` returns as expected
  * There are zero entries in dmesg when this occurs
  * There are zero entries in dmesg when this occurs if the  apparmor profile 
for `chronyd` is placed in complain mode instead of enforce mode
  * We changed the time server from the internal GCP metadata.google.internal 
to the ubuntu time server ntp.ubuntu.com with no change in behaviour

  
  We also noted issues with DNS resolution in snaps like `google-cloud-cli` in 
GCE images. 

  * Disabling apparmor completely for snaps too (`sudo systemctl stop
  snapd.apparmor`) results in no regression and calling the snaps
  returns as expected.

  
  The same issues are present in azure kernel `linux-azure` `6.8.0-1005.5` and 
the -proposed `6.8.0-25.25` generic kernel. 

  This is a release blocker for Noble release

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/chrony/+bug/2061851/+subscriptions


-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 2061851] Re: linux-gcp 6.8.0-1005.5 (+ others) Noble kernel regression iwth new apparmor profiles/features

2024-04-16 Thread Philip Roche
** Also affects: linux-gcp (Ubuntu)
   Importance: Undecided
   Status: New

** Also affects: linux-aws (Ubuntu)
   Importance: Undecided
   Status: New

** Also affects: linux-azure (Ubuntu)
   Importance: Undecided
   Status: New

** Also affects: linux-oracle (Ubuntu)
   Importance: Undecided
   Status: New

** Also affects: linux-ibm (Ubuntu)
   Importance: Undecided
   Status: New

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux-azure in Ubuntu.
https://bugs.launchpad.net/bugs/2061851

Title:
  linux-gcp 6.8.0-1005.5 (+ others) Noble kernel regression iwth new
  apparmor profiles/features

Status in chrony package in Ubuntu:
  New
Status in linux package in Ubuntu:
  New
Status in linux-aws package in Ubuntu:
  New
Status in linux-azure package in Ubuntu:
  New
Status in linux-gcp package in Ubuntu:
  New
Status in linux-ibm package in Ubuntu:
  New
Status in linux-oracle package in Ubuntu:
  New
Status in snapd package in Ubuntu:
  New
Status in chrony source package in Noble:
  New
Status in linux source package in Noble:
  New
Status in linux-aws source package in Noble:
  New
Status in linux-azure source package in Noble:
  New
Status in linux-gcp source package in Noble:
  New
Status in linux-ibm source package in Noble:
  New
Status in linux-oracle source package in Noble:
  New
Status in snapd source package in Noble:
  New

Bug description:
  * Canonical Public Cloud discovered that `chronyc -c sources` now fails with 
`506 Cannot talk to daemon` with the latest kernels. We are seeing this in 
linux-azure and linux-gcp kernels (6.8.0-1005.5)
  * Disabling AppArmor (`sudo systemctl stop apparmor`) completely results in 
no regression and `chronyc -c sources` returns as expected
  * Disabling the apparmor profile for `chronyd` only results in no regression 
and `chronyc -c sources` returns as expected
  * There are zero entries in dmesg when this occurs
  * There are zero entries in dmesg when this occurs if the  apparmor profile 
for `chronyd` is placed in complain mode instead of enforce mode
  * We changed the time server from the internal GCP metadata.google.internal 
to the ubuntu time server ntp.ubuntu.com with no change in behaviour

  
  We also noted issues with DNS resolution in snaps like `google-cloud-cli` in 
GCE images. 

  * Disabling apparmor completely for snaps too (`sudo systemctl stop
  snapd.apparmor`) results in no regression and calling the snaps
  returns as expected.

  
  The same issues are present in azure kernel `linux-azure` `6.8.0-1005.5` and 
the -proposed `6.8.0-25.25` generic kernel. 

  This is a release blocker for Noble release

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/chrony/+bug/2061851/+subscriptions


-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 2061851] Re: linux-gcp 6.8.0-1005.5 (+ others) Noble kernel regression iwth new apparmor profiles/features

2024-04-16 Thread Philip Roche
** Also affects: chrony (Ubuntu)
   Importance: Undecided
   Status: New

** Also affects: snapd (Ubuntu)
   Importance: Undecided
   Status: New

** Also affects: chrony (Ubuntu Noble)
   Importance: Undecided
   Status: New

** Also affects: linux (Ubuntu Noble)
   Importance: Undecided
   Status: New

** Also affects: snapd (Ubuntu Noble)
   Importance: Undecided
   Status: New

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/2061851

Title:
  linux-gcp 6.8.0-1005.5 (+ others) Noble kernel regression iwth new
  apparmor profiles/features

Status in chrony package in Ubuntu:
  New
Status in linux package in Ubuntu:
  New
Status in snapd package in Ubuntu:
  New
Status in chrony source package in Noble:
  New
Status in linux source package in Noble:
  New
Status in snapd source package in Noble:
  New

Bug description:
  * Canonical Public Cloud discovered that `chronyc -c sources` now fails with 
`506 Cannot talk to daemon` with the latest kernels. We are seeing this in 
linux-azure and linux-gcp kernels (6.8.0-1005.5)
  * Disabling AppArmor (`sudo systemctl stop apparmor`) completely results in 
no regression and `chronyc -c sources` returns as expected
  * Disabling the apparmor profile for `chronyd` only results in no regression 
and `chronyc -c sources` returns as expected
  * There are zero entries in dmesg when this occurs
  * There are zero entries in dmesg when this occurs if the  apparmor profile 
for `chronyd` is placed in complain mode instead of enforce mode
  * We changed the time server from the internal GCP metadata.google.internal 
to the ubuntu time server ntp.ubuntu.com with no change in behaviour

  
  We also noted issues with DNS resolution in snaps like `google-cloud-cli` in 
GCE images. 

  * Disabling apparmor completely for snaps too (`sudo systemctl stop
  snapd.apparmor`) results in no regression and calling the snaps
  returns as expected.

  
  The same issues are present in azure kernel `linux-azure` `6.8.0-1005.5` and 
the -proposed `6.8.0-25.25` generic kernel. 

  This is a release blocker for Noble release

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/chrony/+bug/2061851/+subscriptions


-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 2061851] [NEW] linux-gcp 6.8.0-1005.5 (+ others) Noble kernel regression iwth new apparmor profiles/features

2024-04-16 Thread Philip Roche
Public bug reported:

* Canonical Public Cloud discovered that `chronyc -c sources` now fails with 
`506 Cannot talk to daemon` with the latest kernels. We are seeing this in 
linux-azure and linux-gcp kernels (6.8.0-1005.5)
* Disabling AppArmor (`sudo systemctl stop apparmor`) completely results in no 
regression and `chronyc -c sources` returns as expected
* Disabling the apparmor profile for `chronyd` only results in no regression 
and `chronyc -c sources` returns as expected
* There are zero entries in dmesg when this occurs
* There are zero entries in dmesg when this occurs if the  apparmor profile for 
`chronyd` is placed in complain mode instead of enforce mode
* We changed the time server from the internal GCP metadata.google.internal to 
the ubuntu time server ntp.ubuntu.com with no change in behaviour


We also noted issues with DNS resolution in snaps like `google-cloud-cli` in 
GCE images. 

* Disabling apparmor completely for snaps too (`sudo systemctl stop
snapd.apparmor`) results in no regression and calling the snaps returns
as expected.


The same issues are present in azure kernel `linux-azure` `6.8.0-1005.5` and 
the -proposed `6.8.0-25.25` generic kernel. 

This is a release blocker for Noble release

** Affects: chrony (Ubuntu)
 Importance: Undecided
 Status: New

** Affects: linux (Ubuntu)
 Importance: Undecided
 Status: New

** Affects: snapd (Ubuntu)
 Importance: Undecided
 Status: New

** Affects: chrony (Ubuntu Noble)
 Importance: Undecided
 Status: New

** Affects: linux (Ubuntu Noble)
 Importance: Undecided
 Status: New

** Affects: snapd (Ubuntu Noble)
 Importance: Undecided
 Status: New


** Tags: block-proposed block-proposed-noble

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/2061851

Title:
  linux-gcp 6.8.0-1005.5 (+ others) Noble kernel regression iwth new
  apparmor profiles/features

Status in chrony package in Ubuntu:
  New
Status in linux package in Ubuntu:
  New
Status in snapd package in Ubuntu:
  New
Status in chrony source package in Noble:
  New
Status in linux source package in Noble:
  New
Status in snapd source package in Noble:
  New

Bug description:
  * Canonical Public Cloud discovered that `chronyc -c sources` now fails with 
`506 Cannot talk to daemon` with the latest kernels. We are seeing this in 
linux-azure and linux-gcp kernels (6.8.0-1005.5)
  * Disabling AppArmor (`sudo systemctl stop apparmor`) completely results in 
no regression and `chronyc -c sources` returns as expected
  * Disabling the apparmor profile for `chronyd` only results in no regression 
and `chronyc -c sources` returns as expected
  * There are zero entries in dmesg when this occurs
  * There are zero entries in dmesg when this occurs if the  apparmor profile 
for `chronyd` is placed in complain mode instead of enforce mode
  * We changed the time server from the internal GCP metadata.google.internal 
to the ubuntu time server ntp.ubuntu.com with no change in behaviour

  
  We also noted issues with DNS resolution in snaps like `google-cloud-cli` in 
GCE images. 

  * Disabling apparmor completely for snaps too (`sudo systemctl stop
  snapd.apparmor`) results in no regression and calling the snaps
  returns as expected.

  
  The same issues are present in azure kernel `linux-azure` `6.8.0-1005.5` and 
the -proposed `6.8.0-25.25` generic kernel. 

  This is a release blocker for Noble release

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/chrony/+bug/2061851/+subscriptions


-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 2042564] Re: Performance regression in the 5.15 Ubuntu 20.04 kernel compared to 5.4 Ubuntu 20.04 kernel

2023-11-21 Thread Philip Roche
I have reproduced with @amalmostafa's updated script with a separate
disk too. I see no segfault and no EXT4 errors but the regression in
performance is still present but not as great as in my previous tests.

```
### Ubuntu 20.04 with 5.4 kernel and data disk

ubuntu@cloudimg:~$ sudo fio --ioengine=libaio --blocksize=4k --readwrite=write 
--filesize=40G --end_fsync=1 --iodepth=128 --direct=1 --group_reporting 
--numjobs=8 --name=fiojob1 --filename=/dev/sdb
fiojob1: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, 
ioengine=libaio, iodepth=128
...
fio-3.16
Starting 8 processes
Jobs: 8 (f=8): [W(8)][100.0%][w=1162MiB/s][w=298k IOPS][eta 00m:00s]
fiojob1: (groupid=0, jobs=8): err= 0: pid=2391: Tue Nov 21 10:34:57 2023
  write: IOPS=284k, BW=1108MiB/s (1162MB/s)(320GiB/295713msec); 0 zone resets
slat (nsec): min=751, max=115304k, avg=8630.05, stdev=124263.77
clat (nsec): min=391, max=239001k, avg=3598764.23, stdev=2429948.87
 lat (usec): min=72, max=239002, avg=3607.70, stdev=2428.75
clat percentiles (usec):
 |  1.00th=[  668],  5.00th=[ 1434], 10.00th=[ 1778], 20.00th=[ 2212],
 | 30.00th=[ 2573], 40.00th=[ 2900], 50.00th=[ 3261], 60.00th=[ 3654],
 | 70.00th=[ 4080], 80.00th=[ 4686], 90.00th=[ 5669], 95.00th=[ 6587],
 | 99.00th=[ 9110], 99.50th=[10945], 99.90th=[26608], 99.95th=[43779],
 | 99.99th=[83362]
   bw (  MiB/s): min=  667, max= 1341, per=99.98%, avg=1107.88, stdev=13.07, 
samples=4728
   iops: min=170934, max=343430, avg=283618.08, stdev=3346.84, 
samples=4728
  lat (nsec)   : 500=0.01%, 750=0.01%, 1000=0.01%
  lat (usec)   : 4=0.01%, 10=0.01%, 20=0.01%, 50=0.01%, 100=0.01%
  lat (usec)   : 250=0.04%, 500=0.43%, 750=0.66%, 1000=0.51%
  lat (msec)   : 2=13.23%, 4=53.26%, 10=31.18%, 20=0.53%, 50=0.11%
  lat (msec)   : 100=0.03%, 250=0.01%
  cpu  : usr=3.10%, sys=7.36%, ctx=1105263, majf=0, minf=102
  IO depths: 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.1%, 32=0.1%, >=64=100.0%
 submit: 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0%
 complete  : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1%
 issued rwts: total=0,83886080,0,8 short=0,0,0,0 dropped=0,0,0,0
 latency   : target=0, window=0, percentile=100.00%, depth=128

Run status group 0 (all jobs):
  WRITE: bw=1108MiB/s (1162MB/s), 1108MiB/s-1108MiB/s (1162MB/s-1162MB/s), 
io=320GiB (344GB), run=295713-295713msec

Disk stats (read/write):
  sdb: ios=96/33256749, merge=0/50606838, ticks=19/30836895, in_queue=971080, 
util=100.00%
ubuntu@cloudimg:~$ 
ubuntu@cloudimg:~$ uname --kernel-release
5.4.0-164-generic
ubuntu@cloudimg:~$ cat /etc/cloud/build.info 
build_name: server
serial: 20231011
ubuntu@cloudimg:~$ cat /etc/os-release 
NAME="Ubuntu"
VERSION="20.04.6 LTS (Focal Fossa)"
ID=ubuntu
ID_LIKE=debian
PRETTY_NAME="Ubuntu 20.04.6 LTS"
VERSION_ID="20.04"
HOME_URL="https://www.ubuntu.com/;
SUPPORT_URL="https://help.ubuntu.com/;
BUG_REPORT_URL="https://bugs.launchpad.net/ubuntu/;
PRIVACY_POLICY_URL="https://www.ubuntu.com/legal/terms-and-policies/privacy-policy;
VERSION_CODENAME=focal
UBUNTU_CODENAME=focal

### Ubuntu 20.04 with 5.15 kernel and data disk

ubuntu@cloudimg:~$ uname --kernel-release 
5.15.0-89-generic
ubuntu@cloudimg:~$ cat /etc/os-release 
NAME="Ubuntu"
VERSION="20.04.6 LTS (Focal Fossa)"
ID=ubuntu
ID_LIKE=debian
PRETTY_NAME="Ubuntu 20.04.6 LTS"
VERSION_ID="20.04"
HOME_URL="https://www.ubuntu.com/;
SUPPORT_URL="https://help.ubuntu.com/;
BUG_REPORT_URL="https://bugs.launchpad.net/ubuntu/;
PRIVACY_POLICY_URL="https://www.ubuntu.com/legal/terms-and-policies/privacy-policy;
VERSION_CODENAME=focal
UBUNTU_CODENAME=focal
ubuntu@cloudimg:~$ cat /etc/cloud/build.info 
build_name: server
serial: 20231011

ubuntu@cloudimg:~$ sudo fio --ioengine=libaio --blocksize=4k --readwrite=write 
--filesize=40G --end_fsync=1 --iodepth=128 --direct=1 --group_reporting 
--numjobs=8 --name=fiojob1 --filename=/dev/sdb
fiojob1: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, 
ioengine=libaio, iodepth=128
...
fio-3.16
Starting 8 processes
Jobs: 8 (f=8): [W(8)][100.0%][w=1071MiB/s][w=274k IOPS][eta 00m:00s]
fiojob1: (groupid=0, jobs=8): err= 0: pid=1008: Tue Nov 21 12:19:56 2023
  write: IOPS=258k, BW=1007MiB/s (1056MB/s)(320GiB/325284msec); 0 zone resets
slat (nsec): min=931, max=36726k, avg=7936.15, stdev=120427.45
clat (nsec): min=1963, max=155870k, avg=3959799.65, stdev=2129472.51
 lat (usec): min=55, max=155872, avg=3968.10, stdev=2128.87
clat percentiles (usec):
 |  1.00th=[  562],  5.00th=[ 1319], 10.00th=[ 1811], 20.00th=[ 2376],
 | 30.00th=[ 2835], 40.00th=[ 3294], 50.00th=[ 3720], 60.00th=[ 4113],
 | 70.00th=[ 4621], 80.00th=[ 5211], 90.00th=[ 6390], 95.00th=[ 7439],
 | 99.00th=[10159], 99.50th=[11863], 99.90th=[19268], 99.95th=[23462],
 | 99.99th=[36439]
   bw (  KiB/s): min=715896, max=1190887, per=100.00%, avg=1031613.99, 
stdev=9298.50, samples=5200
   iops: 

[Kernel-packages] [Bug 2042564] Re: Performance regression in the 5.15 Ubuntu 20.04 kernel compared to 5.4 Ubuntu 20.04 kernel

2023-11-08 Thread Philip Roche
I am NOT seeing the same on 22.04

```
ubuntu@cloudimg:~$ uname --kernel-release
5.15.0-87-generic
ubuntu@cloudimg:~$ cat /etc/os-release
PRETTY_NAME="Ubuntu 22.04.3 LTS"
NAME="Ubuntu"
VERSION_ID="22.04"
VERSION="22.04.3 LTS (Jammy Jellyfish)"
VERSION_CODENAME=jammy
ID=ubuntu
ID_LIKE=debian
HOME_URL="https://www.ubuntu.com/;
SUPPORT_URL="https://help.ubuntu.com/;
BUG_REPORT_URL="https://bugs.launchpad.net/ubuntu/;
PRIVACY_POLICY_URL="https://www.ubuntu.com/legal/terms-and-policies/privacy-policy;
UBUNTU_CODENAME=jammy
ubuntu@cloudimg:~$ sudo fio --ioengine=libaio --blocksize=4k --readwrite=write 
--filesize=40G --end_fsync=1 --iodepth=128 --direct=1 --group_reporting 
--numjobs=8 --name=fiojob1 --filename=/dev/sda
fiojob1: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, 
ioengine=libaio, iodepth=128
...
fio-3.28
Starting 8 processes
Jobs: 8 (f=8): [W(8)][100.0%][w=1180MiB/s][w=302k IOPS][eta 00m:00s]
fiojob1: (groupid=0, jobs=8): err= 0: pid=2166: Thu Nov  9 07:30:05 2023
  write: IOPS=315k, BW=1230MiB/s (1290MB/s)(320GiB/266343msec); 0 zone resets
slat (nsec): min=662, max=37599k, avg=4806.87, stdev=81368.70
clat (nsec): min=656, max=87622k, avg=3241604.28, stdev=1888945.13
 lat (usec): min=46, max=87623, avg=3246.74, stdev=1888.87
clat percentiles (usec):
 |  1.00th=[  441],  5.00th=[ 1139], 10.00th=[ 1483], 20.00th=[ 1893],
 | 30.00th=[ 2245], 40.00th=[ 2573], 50.00th=[ 2933], 60.00th=[ 3294],
 | 70.00th=[ 3752], 80.00th=[ 4359], 90.00th=[ 5276], 95.00th=[ 6194],
 | 99.00th=[ 9241], 99.50th=[11207], 99.90th=[19530], 99.95th=[24511],
 | 99.99th=[36439]
   bw (  MiB/s): min=  512, max= 1567, per=100.00%, avg=1232.42, stdev=21.42, 
samples=4248
   iops: min=131165, max=401400, avg=315500.53, stdev=5483.45, 
samples=4248
  lat (nsec)   : 750=0.01%, 1000=0.01%
  lat (usec)   : 2=0.01%, 10=0.01%, 20=0.01%, 50=0.01%, 100=0.01%
  lat (usec)   : 250=0.02%, 500=1.41%, 750=0.91%, 1000=1.47%
  lat (msec)   : 2=19.05%, 4=51.85%, 10=24.55%, 20=0.66%, 50=0.09%
  lat (msec)   : 100=0.01%
  cpu  : usr=3.39%, sys=7.68%, ctx=968250, majf=0, minf=108
  IO depths: 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.1%, 32=0.1%, >=64=100.0%
 submit: 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0%
 complete  : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1%
 issued rwts: total=0,83886080,0,0 short=0,0,0,0 dropped=0,0,0,0
 latency   : target=0, window=0, percentile=100.00%, depth=128

Run status group 0 (all jobs):
  WRITE: bw=1230MiB/s (1290MB/s), 1230MiB/s-1230MiB/s (1290MB/s-1290MB/s), 
io=320GiB (344GB), run=266343-266343msec

Disk stats (read/write):
  sda: ios=230/34934643, merge=0/48920965, ticks=54/27582950, 
in_queue=27583025, util=100.00%
```

... but I do see the following in journalctl


```
Nov 09 07:33:03 cloudimg kernel: EXT4-fs warning (device sda1): 
ext4_dirblock_csum_verify:404: inode #80703: comm journalctl: No space for 
directory leaf checksum. Please run e2fsck -D.
Nov 09 07:33:03 cloudimg kernel: EXT4-fs error (device sda1): 
__ext4_find_entry:1689: inode #80703: comm journalctl: checksumming directory 
block 0
Nov 09 07:33:03 cloudimg kernel: Aborting journal on device sda1-8.
Nov 09 07:33:03 cloudimg kernel: EXT4-fs error (device sda1): 
ext4_journal_check_start:83: comm rs:main Q:Reg: Detected aborted journal
Nov 09 07:33:03 cloudimg kernel: EXT4-fs error (device sda1): 
ext4_journal_check_start:83: comm systemd-journal: Detected aborted journal
Nov 09 07:33:03 cloudimg kernel: EXT4-fs (sda1): Remounting filesystem read-only
Nov 09 07:33:03 cloudimg kernel: EXT4-fs warning (device sda1): 
ext4_dirblock_csum_verify:404: inode #80703: comm journalctl: No space for 
directory leaf checksum. Please run e2fsck -D.
Nov 09 07:33:03 cloudimg kernel: EXT4-fs error (device sda1): 
__ext4_find_entry:1689: inode #80703: comm journalctl: checksumming directory 
block 0
Nov 09 07:33:03 cloudimg kernel: EXT4-fs warning (device sda1): 
ext4_dirblock_csum_verify:404: inode #80703: comm journalctl: No space for 
directory leaf checksum. Please run e2fsck -D.
Nov 09 07:33:03 cloudimg kernel: EXT4-fs error (device sda1): 
__ext4_find_entry:1689: inode #80703: comm journalctl: checksumming directory 
block 0
Nov 09 07:33:03 cloudimg kernel: EXT4-fs warning (device sda1): 
ext4_dirblock_csum_verify:404: inode #80703: comm journalctl: No space for 
directory leaf checksum. Please run e2fsck -D.
Nov 09 07:33:03 cloudimg kernel: EXT4-fs error (device sda1): 
__ext4_find_entry:1689: inode #80703: comm journalctl: checksumming directory 
block 0
Nov 09 07:33:03 cloudimg kernel: EXT4-fs warning (device sda1): 
ext4_dirblock_csum_verify:404: inode #80703: comm journalctl: No space for 
directory leaf checksum. Please run e2fsck -D.
Nov 09 07:33:03 cloudimg kernel: EXT4-fs error (device sda1): 
__ext4_find_entry:1689: inode #80703: comm journalctl: checksumming directory 
block 0
Nov 09 07:33:03 cloudimg kernel: 

[Kernel-packages] [Bug 2042564] Re: Performance regression in the 5.15 Ubuntu 20.04 kernel compared to 5.4 Ubuntu 20.04 kernel

2023-11-08 Thread Philip Roche
@kamalmostafa indeed yes. I had missed this.


See below for the output from 20.04 with 5.15 kernel.


```
ubuntu@cloudimg:~$ sudo fio --ioengine=libaio --blocksize=4k --readwrite=write 
--filesize=40G --end_fsync=1 --iodepth=128 --direct=1 --group_reporting 
--numjobs=8 --name=fiojob1 --filename=/dev/sda
fiojob1: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, 
ioengine=libaio, iodepth=128
...
fio-3.16
Starting 8 processes
Jobs: 4 (f=4): [W(1),_(2),W(1),_(1),W(2),_(1)][99.6%][w=212MiB/s][w=54.3k 
IOPS][eta 00m:01s]
fiojob1: (groupid=0, jobs=8): err= 0: pid=697: Thu Nov  9 06:30:01 2023
  write: IOPS=319k, BW=1247MiB/s (1307MB/s)(320GiB/262803msec); 0 zone resets
slat (nsec): min=652, max=55305k, avg=4699.55, stdev=76242.73
clat (usec): min=5, max=121658, avg=3187.92, stdev=1886.86
 lat (usec): min=34, max=121660, avg=3192.96, stdev=1886.68
clat percentiles (usec):
 |  1.00th=[  506],  5.00th=[ 1254], 10.00th=[ 1565], 20.00th=[ 1909],
 | 30.00th=[ 2212], 40.00th=[ 2507], 50.00th=[ 2835], 60.00th=[ 3228],
 | 70.00th=[ 3687], 80.00th=[ 4228], 90.00th=[ 5145], 95.00th=[ 5997],
 | 99.00th=[ 8717], 99.50th=[10683], 99.90th=[18744], 99.95th=[25822],
 | 99.99th=[55837]
   bw (  MiB/s): min=  353, max= 1690, per=100.00%, avg=1251.19, stdev=22.89, 
samples=4182
   iops: min=90615, max=432642, avg=320303.28, stdev=5860.14, 
samples=4182
  lat (usec)   : 10=0.01%, 20=0.01%, 50=0.01%, 100=0.01%, 250=0.02%
  lat (usec)   : 500=0.95%, 750=0.55%, 1000=1.01%
  lat (msec)   : 2=20.21%, 4=53.10%, 10=23.55%, 20=0.52%, 50=0.07%
  lat (msec)   : 100=0.01%, 250=0.01%
  cpu  : usr=3.42%, sys=7.71%, ctx=960980, majf=0, minf=86
  IO depths: 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.1%, 32=0.1%, >=64=100.0%
 submit: 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0%
 complete  : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1%
 issued rwts: total=0,83886080,0,8 short=0,0,0,0 dropped=0,0,0,0
 latency   : target=0, window=0, percentile=100.00%, depth=128

Run status group 0 (all jobs):
  WRITE: bw=1247MiB/s (1307MB/s), 1247MiB/s-1247MiB/s (1307MB/s-1307MB/s), 
io=320GiB (344GB), run=262803-262803msec

Disk stats (read/write):
  sda: ios=251/31989123, merge=0/51876731, ticks=482/27581400, 
in_queue=27581891, util=100.00%
Segmentation fault
ubuntu@cloudimg:~$ uname --kernel-release
5.15.0-88-generic
ubuntu@cloudimg:~$ cat /etc/os-release 
NAME="Ubuntu"
VERSION="20.04.6 LTS (Focal Fossa)"
ID=ubuntu
ID_LIKE=debian
PRETTY_NAME="Ubuntu 20.04.6 LTS"
VERSION_ID="20.04"
HOME_URL="https://www.ubuntu.com/;
SUPPORT_URL="https://help.ubuntu.com/;
BUG_REPORT_URL="https://bugs.launchpad.net/ubuntu/;
PRIVACY_POLICY_URL="https://www.ubuntu.com/legal/terms-and-policies/privacy-policy;
VERSION_CODENAME=focal
UBUNTU_CODENAME=focal
ubuntu@cloudimg:~$ 
```

I also see the same details in `journalctl`

```
Nov 09 06:25:38 cloudimg sudo[692]:   ubuntu : TTY=pts/0 ; PWD=/home/ubuntu ; 
USER=root ; COMMAND=/usr/bin/fio --ioengine=libaio --blocksize=4k 
--readwrite=write --filesize=40G --end_fsync=1 --iodepth=128 --direct=1 
--group_reporting --numjobs=8 --name=fiojob1 --filename=/dev/sda
Nov 09 06:25:38 cloudimg sudo[692]: pam_unix(sudo:session): session opened for 
user root by ubuntu(uid=0)
Nov 09 06:30:01 cloudimg kernel: fio[693]: segfault at 7f62e5d78595 ip 
7f62fa672a50 sp 7ffeaecfd2a8 error 6 in 
libglusterfs.so.0.0.1[7f62fa5b+c3000]
Nov 09 06:30:01 cloudimg kernel: Code: f0 0e d0 dd 14 e5 db 95 d9 14 73 00 00 
00 00 00 00 00 c2 11 ee 01 00 00 00 00 00 3a 21 70 00 00 00 00 4d f9 e0 18 e7 
be 9b 09 <29> 9f 75 66 6c eb 9a 03 e5 33 3a bd 32 2a 9c 0e 7c c6 b4 3d a9 26
Nov 09 06:30:01 cloudimg kernel: Core dump to |/usr/share/apport/apport pipe 
failed
Nov 09 06:30:01 cloudimg sudo[692]: pam_unix(sudo:session): session closed for 
user root
Nov 09 06:30:01 cloudimg kernel: Core dump to |/usr/share/apport/apport pipe 
failed
Nov 09 06:39:19 cloudimg systemd[1]: Starting Cleanup of Temporary 
Directories...
Nov 09 06:39:19 cloudimg kernel: EXT4-fs warning (device sda1): 
ext4_dirblock_csum_verify:404: inode #108: comm systemd-tmpfile: No space for 
directory leaf checksum. Please run e2fsck -D.
Nov 09 06:39:19 cloudimg kernel: EXT4-fs error (device sda1): 
htree_dirblock_to_tree:1080: inode #108: comm systemd-tmpfile: Directory block 
failed checksum
Nov 09 06:39:19 cloudimg systemd[1]: systemd-tmpfiles-clean.service: Succeeded.
```

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/2042564

Title:
  Performance regression in the 5.15 Ubuntu 20.04 kernel compared to 5.4
  Ubuntu 20.04 kernel

Status in linux package in Ubuntu:
  New
Status in linux source package in Focal:
  New

Bug description:
  We in the Canonical Public Cloud team have received report from our
  colleagues in Google regarding a potential performance 

[Kernel-packages] [Bug 2042564] Re: Performance regression in the 5.15 Ubuntu 20.04 kernel compared to 5.4 Ubuntu 20.04 kernel

2023-11-08 Thread Philip Roche
Providing exact reproducer steps using Qemu locally -

Launch script:
https://gist.github.com/philroche/8242106415ef35b446d7e625b6d60c90 and
cloud image I used for testing @ http://cloud-
images.ubuntu.com/minimal/releases/focal/release/


```
# Download the VM launch script
wget --output-document=launch-qcow2-image-qemu-40G.sh 
https://gist.githubusercontent.com/philroche/8242106415ef35b446d7e625b6d60c90/raw/4a338b92301b6e08608e9345f85a50452ca5fa21/launch-qcow2-image-qemu-40G.sh

chmod +x launch-qcow2-image-qemu-40G.sh

# Download latest  image
wget --output-document=ubuntu-20.04-minimal-cloudimg-amd64.img 
http://cloud-images.ubuntu.com/minimal/releases/focal/release/ubuntu-20.04-minimal-cloudimg-amd64.img

# launch the image - note that this provisions a 40GB disk.
./launch-qcow2-image-qemu-40G.sh --password passw0rd --image 
./20231102-ubuntu-20.04-minimal-cloudimg-amd64.img 

# install fio
sshpass -p passw0rd ssh -p  -o UserKnownHostsFile=/dev/null -o 
CheckHostIP=no -o StrictHostKeyChecking=no ubuntu@127.0.0.1 -- sudo apt-get 
update

sshpass -p passw0rd ssh -p  -o UserKnownHostsFile=/dev/null -o
CheckHostIP=no -o StrictHostKeyChecking=no ubuntu@127.0.0.1 -- sudo apt-
get install --assume-yes fio

# run the synthetic test and gather information
sshpass -p passw0rd ssh -p  -o UserKnownHostsFile=/dev/null -o 
CheckHostIP=no -o StrictHostKeyChecking=no ubuntu@127.0.0.1 -- sudo fio 
--ioengine=libaio --blocksize=4k --readwrite=write --filesize=40G --end_fsync=1 
--iodepth=128 --direct=1 --group_reporting --numjobs=8 --name=fiojob1 
--filename=/dev/sda

# kill VM (Use 'Ctrl-a x' key combination to exit emulator )

# start new VM to upgrade kernel and run the synthetic test again
# launch the image - note that this provisions a 40GB disk.
./launch-qcow2-image-qemu-40G.sh --password passw0rd --image 
./20231102-ubuntu-20.04-minimal-cloudimg-amd64.img 

# install fio
sshpass -p passw0rd ssh -p  -o UserKnownHostsFile=/dev/null -o 
CheckHostIP=no -o StrictHostKeyChecking=no ubuntu@127.0.0.1 -- sudo apt-get 
update

sshpass -p passw0rd ssh -p  -o UserKnownHostsFile=/dev/null -o
CheckHostIP=no -o StrictHostKeyChecking=no ubuntu@127.0.0.1 -- sudo apt-
get install --assume-yes fio

# upgrade the kernel
sshpass -p passw0rd ssh -p  -o UserKnownHostsFile=/dev/null -o 
CheckHostIP=no -o StrictHostKeyChecking=no ubuntu@127.0.0.1 -- sudo apt-get 
install --assume-yes linux-generic-hwe-20.04

# wait until VM has rebooted
sshpass -p passw0rd ssh -p  -o UserKnownHostsFile=/dev/null -o 
CheckHostIP=no -o StrictHostKeyChecking=no ubuntu@127.0.0.1 -- sudo reboot

# run the synthetic test again and gather the information
sshpass -p passw0rd ssh -p  -o UserKnownHostsFile=/dev/null -o 
CheckHostIP=no -o StrictHostKeyChecking=no ubuntu@127.0.0.1 -- sudo fio 
--ioengine=libaio --blocksize=4k --readwrite=write --filesize=40G --end_fsync=1 
--iodepth=128 --direct=1 --group_reporting --numjobs=8 --name=fiojob1 
--filename=/dev/sda

```

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/2042564

Title:
  Performance regression in the 5.15 Ubuntu 20.04 kernel compared to 5.4
  Ubuntu 20.04 kernel

Status in linux package in Ubuntu:
  New
Status in linux source package in Focal:
  New

Bug description:
  We in the Canonical Public Cloud team have received report from our
  colleagues in Google regarding a potential performance regression with
  the 5.15 kernel vs the 5.4 kernel on ubuntu 20.04. Their test were
  performed using the linux-gkeop and linux-gkeop-5.15 kernels.

  I have verified with the generic Ubuntu 20.04 5.4 linux-generic and
  the Ubuntu 20.04 5.15 linux-generic-hwe-20.04 kernels.

  The tests were run using `fio`

  fio commands:

  * 4k initwrite: `fio --ioengine=libaio --blocksize=4k --readwrite=write 
--filesize=40G --end_fsync=1 --iodepth=128 --direct=1 --group_reporting 
--numjobs=8 --name=fiojob1 --filename=/dev/sdc`
  * 4k overwrite: `fio --ioengine=libaio --blocksize=4k --readwrite=write 
--filesize=40G --end_fsync=1 --iodepth=128 --direct=1 --group_reporting 
--numjobs=8 --name=fiojob1 --filename=/dev/sdc`

  
  My reproducer was to launch an Ubuntu 20.04 cloud image locally with qemu the 
results are below:

  Using 5.4 kernel

  ```
  ubuntu@cloudimg:~$ uname --kernel-release
  5.4.0-164-generic

  ubuntu@cloudimg:~$ sudo fio --ioengine=libaio --blocksize=4k 
--readwrite=write --filesize=40G --end_fsync=1 --iodepth=128 --direct=1 
--group_reporting --numjobs=8 --name=fiojob1 --filename=/dev/sda
  fiojob1: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 
4096B-4096B, ioengine=libaio, iodepth=128
  ...
  fio-3.16
  Starting 8 processes
  Jobs: 8 (f=8): [W(8)][99.6%][w=925MiB/s][w=237k IOPS][eta 00m:01s] 
  fiojob1: (groupid=0, jobs=8): err= 0: pid=2443: Thu Nov  2 09:15:22 2023
write: IOPS=317k, BW=1237MiB/s (1297MB/s)(320GiB/264837msec); 0 zone resets
  

[Kernel-packages] [Bug 2042564] Re: Performance regression in the 5.15 Ubuntu 20.04 kernel compared to 5.4 Ubuntu 20.04 kernel

2023-11-07 Thread Philip Roche
Google have provided a non synthetic `fio` impact

> Performance was severally degraded when accessing Persistent Volumes
provided by Portworx/PureStorage.

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/2042564

Title:
  Performance regression in the 5.15 Ubuntu 20.04 kernel compared to 5.4
  Ubuntu 20.04 kernel

Status in linux package in Ubuntu:
  New
Status in linux source package in Focal:
  New

Bug description:
  We in the Canonical Public Cloud team have received report from our
  colleagues in Google regarding a potential performance regression with
  the 5.15 kernel vs the 5.4 kernel on ubuntu 20.04. Their test were
  performed using the linux-gkeop and linux-gkeop-5.15 kernels.

  I have verified with the generic Ubuntu 20.04 5.4 linux-generic and
  the Ubuntu 20.04 5.15 linux-generic-hwe-20.04 kernels.

  The tests were run using `fio`

  fio commands:

  * 4k initwrite: `fio --ioengine=libaio --blocksize=4k --readwrite=write 
--filesize=40G --end_fsync=1 --iodepth=128 --direct=1 --group_reporting 
--numjobs=8 --name=fiojob1 --filename=/dev/sdc`
  * 4k overwrite: `fio --ioengine=libaio --blocksize=4k --readwrite=write 
--filesize=40G --end_fsync=1 --iodepth=128 --direct=1 --group_reporting 
--numjobs=8 --name=fiojob1 --filename=/dev/sdc`

  
  My reproducer was to launch an Ubuntu 20.04 cloud image locally with qemu the 
results are below:

  Using 5.4 kernel

  ```
  ubuntu@cloudimg:~$ uname --kernel-release
  5.4.0-164-generic

  ubuntu@cloudimg:~$ sudo fio --ioengine=libaio --blocksize=4k 
--readwrite=write --filesize=40G --end_fsync=1 --iodepth=128 --direct=1 
--group_reporting --numjobs=8 --name=fiojob1 --filename=/dev/sda
  fiojob1: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 
4096B-4096B, ioengine=libaio, iodepth=128
  ...
  fio-3.16
  Starting 8 processes
  Jobs: 8 (f=8): [W(8)][99.6%][w=925MiB/s][w=237k IOPS][eta 00m:01s] 
  fiojob1: (groupid=0, jobs=8): err= 0: pid=2443: Thu Nov  2 09:15:22 2023
write: IOPS=317k, BW=1237MiB/s (1297MB/s)(320GiB/264837msec); 0 zone resets
  slat (nsec): min=628, max=37820k, avg=7207.71, stdev=101058.61
  clat (nsec): min=457, max=56099k, avg=340.45, stdev=1707823.38
   lat (usec): min=23, max=56100, avg=3229.78, stdev=1705.80
  clat percentiles (usec):
   |  1.00th=[  775],  5.00th=[ 1352], 10.00th=[ 1647], 20.00th=[ 2024],
   | 30.00th=[ 2343], 40.00th=[ 2638], 50.00th=[ 2933], 60.00th=[ 3261],
   | 70.00th=[ 3654], 80.00th=[ 4146], 90.00th=[ 5014], 95.00th=[ 5932],
   | 99.00th=[ 8979], 99.50th=[10945], 99.90th=[18220], 99.95th=[22676],
   | 99.99th=[32113]
 bw (  MiB/s): min=  524, max= 1665, per=100.00%, avg=1237.72, stdev=20.42, 
samples=4232
 iops: min=134308, max=426326, avg=316855.16, stdev=5227.36, 
samples=4232
lat (nsec)   : 500=0.01%, 750=0.01%, 1000=0.01%
lat (usec)   : 4=0.01%, 10=0.01%, 20=0.01%, 50=0.01%, 100=0.01%
lat (usec)   : 250=0.05%, 500=0.54%, 750=0.37%, 1000=0.93%
lat (msec)   : 2=17.40%, 4=58.02%, 10=22.01%, 20=0.60%, 50=0.07%
lat (msec)   : 100=0.01%
cpu  : usr=3.29%, sys=7.45%, ctx=1262621, majf=0, minf=103
IO depths: 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.1%, 32=0.1%, >=64=100.0%
   submit: 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, 
>=64=0.0%
   complete  : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, 
>=64=0.1%
   issued rwts: total=0,83886080,0,8 short=0,0,0,0 dropped=0,0,0,0
   latency   : target=0, window=0, percentile=100.00%, depth=128

  Run status group 0 (all jobs):
WRITE: bw=1237MiB/s (1297MB/s), 1237MiB/s-1237MiB/s (1297MB/s-1297MB/s), 
io=320GiB (344GB), run=264837-264837msec

  Disk stats (read/write):
sda: ios=36/32868891, merge=0/50979424, ticks=5/27498602, in_queue=1183124, 
util=100.00%
  ```

  
  After upgrading to linux-generic-hwe-20.04 kernel and rebooting

  ```
  ubuntu@cloudimg:~$ uname --kernel-release
  5.15.0-88-generic

  ubuntu@cloudimg:~$ sudo fio --ioengine=libaio --blocksize=4k 
--readwrite=write --filesize=40G --end_fsync=1 --iodepth=128 --direct=1 
--group_reporting --numjobs=8 --name=fiojob1 --filename=/dev/sda
  fiojob1: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 
4096B-4096B, ioengine=libaio, iodepth=128
  ...
  fio-3.16
  Starting 8 processes
  Jobs: 1 (f=1): [_(7),W(1)][100.0%][w=410MiB/s][w=105k IOPS][eta 00m:00s]
  fiojob1: (groupid=0, jobs=8): err= 0: pid=1438: Thu Nov  2 09:46:49 2023
write: IOPS=155k, BW=605MiB/s (634MB/s)(320GiB/541949msec); 0 zone resets
  slat (nsec): min=660, max=325426k, avg=10351.04, stdev=232438.50
  clat (nsec): min=1100, max=782743k, avg=6595008.67, stdev=6290570.04
   lat (usec): min=86, max=782748, avg=6606.08, stdev=6294.03
  clat percentiles (usec):
   |  1.00th=[   914],  5.00th=[  2180], 10.00th=[  2802], 20.00th=[  3556],
   | 30.00th=[  

[Kernel-packages] [Bug 2042564] Re: Performance regression in the 5.15 Ubuntu 20.04 kernel compared to 5.4 Ubuntu 20.04 kernel

2023-11-02 Thread Philip Roche
For reference, I also tried on a 22.04 cloud image with 5.15 kernel

```
ubuntu@cloudimg:~$ sudo fio --ioengine=libaio --blocksize=4k --readwrite=write 
--filesize=40G --end_fsync=1 --iodepth=128 --direct=1 --group_reporting 
--numjobs=8 --name=fiojob1 --filename=/dev/sda
fiojob1: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, 
ioengine=libaio, iodepth=128
...
fio-3.28
Starting 8 processes
Jobs: 1 (f=1): [_(5),W(1),_(2)][100.0%][w=311MiB/s][w=79.7k IOPS][eta 00m:00s]
fiojob1: (groupid=0, jobs=8): err= 0: pid=2198: Thu Nov  2 13:42:07 2023
  write: IOPS=244k, BW=952MiB/s (998MB/s)(320GiB/344252msec); 0 zone resets
slat (nsec): min=655, max=86425k, avg=6303.33, stdev=121079.62
clat (usec): min=5, max=124812, avg=4182.79, stdev=3154.25
 lat (usec): min=61, max=124814, avg=4189.52, stdev=3155.57
clat percentiles (usec):
 |  1.00th=[  553],  5.00th=[ 1319], 10.00th=[ 1713], 20.00th=[ 2180],
 | 30.00th=[ 2638], 40.00th=[ 3064], 50.00th=[ 3523], 60.00th=[ 4015],
 | 70.00th=[ 4621], 80.00th=[ 5538], 90.00th=[ 7177], 95.00th=[ 9110],
 | 99.00th=[15270], 99.50th=[19792], 99.90th=[34866], 99.95th=[42206],
 | 99.99th=[64226]
   bw (  KiB/s): min=197710, max=1710611, per=100.00%, avg=978380.45, 
stdev=38019.38, samples=5483
   iops: min=49424, max=427652, avg=244594.42, stdev=9504.89, 
samples=5483
  lat (usec)   : 10=0.01%, 20=0.01%, 50=0.01%, 100=0.01%, 250=0.01%
  lat (usec)   : 500=0.79%, 750=0.82%, 1000=0.95%
  lat (msec)   : 2=13.06%, 4=44.04%, 10=36.63%, 20=3.21%, 50=0.45%
  lat (msec)   : 100=0.03%, 250=0.01%
  cpu  : usr=3.34%, sys=7.56%, ctx=978696, majf=0, minf=122
  IO depths: 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.1%, 32=0.1%, >=64=100.0%
 submit: 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0%
 complete  : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1%
 issued rwts: total=0,83886080,0,0 short=0,0,0,0 dropped=0,0,0,0
 latency   : target=0, window=0, percentile=100.00%, depth=128

Run status group 0 (all jobs):
  WRITE: bw=952MiB/s (998MB/s), 952MiB/s-952MiB/s (998MB/s-998MB/s), io=320GiB 
(344GB), run=344252-344252msec

Disk stats (read/write):
  sda: ios=336/33974149, merge=0/49890543, ticks=201/35842485, 
in_queue=35842765, util=100.00%
ubuntu@cloudimg:~$ uname --kernel-release
5.15.0-87-generic
```

This shows bw=952MiB/s

Summary:

Ubuntu 22.04 5.15 kernel `bw=952MiB/s`
Ubuntu 20.04 5.4 kernel `bw=1237MiB/s`
Ubuntu 20.04 5.15 kernel `bw=605MiB/s`

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/2042564

Title:
  Performance regression in the 5.15 Ubuntu 20.04 kernel compared to 5.4
  Ubuntu 20.04 kernel

Status in linux package in Ubuntu:
  New
Status in linux source package in Focal:
  New

Bug description:
  We in the Canonical Public Cloud team have received report from our
  colleagues in Google regarding a potential performance regression with
  the 5.15 kernel vs the 5.4 kernel on ubuntu 20.04. Their test were
  performed using the linux-gkeop and linux-gkeop-5.15 kernels.

  I have verified with the generic Ubuntu 20.04 5.4 linux-generic and
  the Ubuntu 20.04 5.15 linux-generic-hwe-20.04 kernels.

  The tests were run using `fio`

  fio commands:

  * 4k initwrite: `fio --ioengine=libaio --blocksize=4k --readwrite=write 
--filesize=40G --end_fsync=1 --iodepth=128 --direct=1 --group_reporting 
--numjobs=8 --name=fiojob1 --filename=/dev/sdc`
  * 4k overwrite: `fio --ioengine=libaio --blocksize=4k --readwrite=write 
--filesize=40G --end_fsync=1 --iodepth=128 --direct=1 --group_reporting 
--numjobs=8 --name=fiojob1 --filename=/dev/sdc`

  
  My reproducer was to launch an Ubuntu 20.04 cloud image locally with qemu the 
results are below:

  Using 5.4 kernel

  ```
  ubuntu@cloudimg:~$ uname --kernel-release
  5.4.0-164-generic

  ubuntu@cloudimg:~$ sudo fio --ioengine=libaio --blocksize=4k 
--readwrite=write --filesize=40G --end_fsync=1 --iodepth=128 --direct=1 
--group_reporting --numjobs=8 --name=fiojob1 --filename=/dev/sda
  fiojob1: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 
4096B-4096B, ioengine=libaio, iodepth=128
  ...
  fio-3.16
  Starting 8 processes
  Jobs: 8 (f=8): [W(8)][99.6%][w=925MiB/s][w=237k IOPS][eta 00m:01s] 
  fiojob1: (groupid=0, jobs=8): err= 0: pid=2443: Thu Nov  2 09:15:22 2023
write: IOPS=317k, BW=1237MiB/s (1297MB/s)(320GiB/264837msec); 0 zone resets
  slat (nsec): min=628, max=37820k, avg=7207.71, stdev=101058.61
  clat (nsec): min=457, max=56099k, avg=340.45, stdev=1707823.38
   lat (usec): min=23, max=56100, avg=3229.78, stdev=1705.80
  clat percentiles (usec):
   |  1.00th=[  775],  5.00th=[ 1352], 10.00th=[ 1647], 20.00th=[ 2024],
   | 30.00th=[ 2343], 40.00th=[ 2638], 50.00th=[ 2933], 60.00th=[ 3261],
   | 70.00th=[ 3654], 80.00th=[ 4146], 90.00th=[ 5014], 95.00th=[ 5932],
   | 

[Kernel-packages] [Bug 2042564] [NEW] Performance regression in the 5.15 Ubuntu 20.04 kernel compared to 5.4 Ubuntu 20.04 kernel

2023-11-02 Thread Philip Roche
Public bug reported:

We in the Canonical Public Cloud team have received report from our
colleagues in Google regarding a potential performance regression with
the 5.15 kernel vs the 5.4 kernel on ubuntu 20.04. Their test were
performed using the linux-gkeop and linux-gkeop-5.15 kernels.

I have verified with the generic Ubuntu 20.04 5.4 linux-generic and the
Ubuntu 20.04 5.15 linux-generic-hwe-20.04 kernels.

The tests were run using `fio`

fio commands:

* 4k initwrite: `fio --ioengine=libaio --blocksize=4k --readwrite=write 
--filesize=40G --end_fsync=1 --iodepth=128 --direct=1 --group_reporting 
--numjobs=8 --name=fiojob1 --filename=/dev/sdc`
* 4k overwrite: `fio --ioengine=libaio --blocksize=4k --readwrite=write 
--filesize=40G --end_fsync=1 --iodepth=128 --direct=1 --group_reporting 
--numjobs=8 --name=fiojob1 --filename=/dev/sdc`


My reproducer was to launch an Ubuntu 20.04 cloud image locally with qemu the 
results are below:

Using 5.4 kernel

```
ubuntu@cloudimg:~$ uname --kernel-release
5.4.0-164-generic

ubuntu@cloudimg:~$ sudo fio --ioengine=libaio --blocksize=4k --readwrite=write 
--filesize=40G --end_fsync=1 --iodepth=128 --direct=1 --group_reporting 
--numjobs=8 --name=fiojob1 --filename=/dev/sda
fiojob1: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, 
ioengine=libaio, iodepth=128
...
fio-3.16
Starting 8 processes
Jobs: 8 (f=8): [W(8)][99.6%][w=925MiB/s][w=237k IOPS][eta 00m:01s] 
fiojob1: (groupid=0, jobs=8): err= 0: pid=2443: Thu Nov  2 09:15:22 2023
  write: IOPS=317k, BW=1237MiB/s (1297MB/s)(320GiB/264837msec); 0 zone resets
slat (nsec): min=628, max=37820k, avg=7207.71, stdev=101058.61
clat (nsec): min=457, max=56099k, avg=340.45, stdev=1707823.38
 lat (usec): min=23, max=56100, avg=3229.78, stdev=1705.80
clat percentiles (usec):
 |  1.00th=[  775],  5.00th=[ 1352], 10.00th=[ 1647], 20.00th=[ 2024],
 | 30.00th=[ 2343], 40.00th=[ 2638], 50.00th=[ 2933], 60.00th=[ 3261],
 | 70.00th=[ 3654], 80.00th=[ 4146], 90.00th=[ 5014], 95.00th=[ 5932],
 | 99.00th=[ 8979], 99.50th=[10945], 99.90th=[18220], 99.95th=[22676],
 | 99.99th=[32113]
   bw (  MiB/s): min=  524, max= 1665, per=100.00%, avg=1237.72, stdev=20.42, 
samples=4232
   iops: min=134308, max=426326, avg=316855.16, stdev=5227.36, 
samples=4232
  lat (nsec)   : 500=0.01%, 750=0.01%, 1000=0.01%
  lat (usec)   : 4=0.01%, 10=0.01%, 20=0.01%, 50=0.01%, 100=0.01%
  lat (usec)   : 250=0.05%, 500=0.54%, 750=0.37%, 1000=0.93%
  lat (msec)   : 2=17.40%, 4=58.02%, 10=22.01%, 20=0.60%, 50=0.07%
  lat (msec)   : 100=0.01%
  cpu  : usr=3.29%, sys=7.45%, ctx=1262621, majf=0, minf=103
  IO depths: 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.1%, 32=0.1%, >=64=100.0%
 submit: 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0%
 complete  : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1%
 issued rwts: total=0,83886080,0,8 short=0,0,0,0 dropped=0,0,0,0
 latency   : target=0, window=0, percentile=100.00%, depth=128

Run status group 0 (all jobs):
  WRITE: bw=1237MiB/s (1297MB/s), 1237MiB/s-1237MiB/s (1297MB/s-1297MB/s), 
io=320GiB (344GB), run=264837-264837msec

Disk stats (read/write):
  sda: ios=36/32868891, merge=0/50979424, ticks=5/27498602, in_queue=1183124, 
util=100.00%
```


After upgrading to linux-generic-hwe-20.04 kernel and rebooting

```
ubuntu@cloudimg:~$ uname --kernel-release
5.15.0-88-generic

ubuntu@cloudimg:~$ sudo fio --ioengine=libaio --blocksize=4k --readwrite=write 
--filesize=40G --end_fsync=1 --iodepth=128 --direct=1 --group_reporting 
--numjobs=8 --name=fiojob1 --filename=/dev/sda
fiojob1: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, 
ioengine=libaio, iodepth=128
...
fio-3.16
Starting 8 processes
Jobs: 1 (f=1): [_(7),W(1)][100.0%][w=410MiB/s][w=105k IOPS][eta 00m:00s]
fiojob1: (groupid=0, jobs=8): err= 0: pid=1438: Thu Nov  2 09:46:49 2023
  write: IOPS=155k, BW=605MiB/s (634MB/s)(320GiB/541949msec); 0 zone resets
slat (nsec): min=660, max=325426k, avg=10351.04, stdev=232438.50
clat (nsec): min=1100, max=782743k, avg=6595008.67, stdev=6290570.04
 lat (usec): min=86, max=782748, avg=6606.08, stdev=6294.03
clat percentiles (usec):
 |  1.00th=[   914],  5.00th=[  2180], 10.00th=[  2802], 20.00th=[  3556],
 | 30.00th=[  4178], 40.00th=[  4817], 50.00th=[  5538], 60.00th=[  6259],
 | 70.00th=[  7177], 80.00th=[  8455], 90.00th=[ 10683], 95.00th=[ 13566],
 | 99.00th=[ 26870], 99.50th=[ 34866], 99.90th=[ 63177], 99.95th=[ 80217],
 | 99.99th=[145753]
   bw (  KiB/s): min=39968, max=1683451, per=100.00%, avg=619292.10, 
stdev=26377.19, samples=8656
   iops: min= 9990, max=420862, avg=154822.58, stdev=6594.34, 
samples=8656
  lat (usec)   : 2=0.01%, 4=0.01%, 10=0.01%, 20=0.01%, 50=0.01%
  lat (usec)   : 100=0.01%, 250=0.01%, 500=0.05%, 750=0.48%, 1000=0.65%
  lat (msec)   : 2=2.79%, 4=23.00%, 10=60.93%, 20=10.08%, 50=1.83%
  lat (msec)   : 

[Kernel-packages] [Bug 2032933] Re: Mantic (23.10) minimal images increase in memory consumption, port usage and processes running

2023-10-10 Thread Philip Roche
https://bugs.launchpad.net/cloud-images/+bug/2038894 is a related bug to
track specifically the introduction of listening port 5353

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/2032933

Title:
  Mantic (23.10) minimal images increase in memory consumption, port
  usage and processes running

Status in cloud-images:
  Confirmed
Status in linux package in Ubuntu:
  Confirmed

Bug description:
  The Mantic (Ubuntu 23.10) download/qcow2 images available @ 
https://cloud-images.ubuntu.com/minimal/
  are undergoing some big changes prior to 23.10 release in October.

  This is a devel release so this is the perfect time to be making these
  changes but we are noticing some changes that were not expected.

  This bug is to track the unexpected changes and discuss/resolve these.

  The changes that have been made to mantic minimal:

  * Move to the linux-generic kernel from the linux-kvm kernel
    * This also involved removal of the virtio-blk driver, which is the default 
for QEMU and OpenStack, but this is being restored in an upcoming 6.5 mantic 
kernel and is being trakced @ 
https://bugs.launchpad.net/ubuntu/+source/linux/+bug/2030745
  * Move to using minimal-cloud seed - see 
https://ubuntu-archive-team.ubuntu.com/seeds/ubuntu.mantic/cloud-minimal
  * No longer installing Recommends packages
* This is during image build only and will not affect any subsequent 
package installs
  * No initramfs fallback for boot - only initramfsless boot

  The latest mantic minimal images are available @ http://cloud-
  images.ubuntu.com/minimal/daily/mantic/ and are also available in the
  public clouds.

  A package name manifest diff can be seen @
  https://pastebin.ubuntu.com/p/rRd6STnNmK/

  We have had reports of higher memory usage on an idle system, higher
  number of ports open on an idle system and higher number of process
  running on a idle system.

  To help with debugging I have built and uploaded the following images
  and package manifests to
  https://people.canonical.com/~philroche/20230824-manticl-minimal-
  LP2032933/

  * 
20230618-before-kernel-change-before-seed-change-mantic-minimal-cloudimg-amd64
    * Before kernel change and before seed change
  * 
20230824-after-kernel-change-before-seed-change-mantic-minimal-cloudimg-amd64
    * After kernel change and before seed change
  * 
20230821.1-after-kernel-change-after-seed-change-mantic-minimal-cloudimg-amd64
    * After kernel change and after seed change

To manage notifications about this bug go to:
https://bugs.launchpad.net/cloud-images/+bug/2032933/+subscriptions


-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 2038567] Re: Disable restricting unprivileged change_profile by default, due to LXD latest/stable not yet compatible with this new apparmor feature

2023-10-09 Thread Philip Roche
cloud minimized and non minimized images have now been tested with
6.5.0-9 kernel from -proposed and pass our lxd-start-stop test suite
which was failing and which is the test suite which prompted this whole
thread. +1

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/2038567

Title:
  Disable restricting unprivileged change_profile by default, due to LXD
  latest/stable not yet compatible with this new apparmor feature

Status in Release Notes for Ubuntu:
  New
Status in apparmor package in Ubuntu:
  New
Status in linux package in Ubuntu:
  Fix Committed
Status in lxd package in Ubuntu:
  Triaged
Status in snapd package in Ubuntu:
  New

Bug description:
  Following upgrade to 6.5.0-7 kernel in mantic cloud images we are
  seeing a regression in our cloud image tests. The test runs the
  following:

  ```
  lxd init --auto --storage-backend dir
  lxc launch ubuntu-daily:mantic mantic
  lxc info mantic
  lxc exec mantic -- cloud-init status --wait
  ```

  The `lxc exec mantic -- cloud-init status --wait` times out after 240s
  and will fail our test as a result.

  I have been able to replicate in a local VM

  ```
  wget 
http://cloud-images.ubuntu.com/mantic/20231005/mantic-server-cloudimg-amd64.img 
  wget --output-document=launch-qcow2-image-qemu.sh 
https://gist.githubusercontent.com/philroche/14c241c086a5730481e24178b654268f/raw/7af95cd4dfc8e1d0600e6118803d2c866765714e/gistfile1.txt
 
  chmod +x launch-qcow2-image-qemu.sh 

  ./launch-qcow2-image-qemu.sh --password passw0rd --image 
./mantic-server-cloudimg-amd64.img 
  cat < "./reproducer.sh"
  #!/bin/bash -eux
  lxd init --auto --storage-backend dir
  lxc launch ubuntu-daily:mantic mantic
  lxc info mantic
  lxc exec mantic -- cloud-init status --wait
  EOF
  chmod +x ./reproducer.sh
  sshpass -p passw0rd scp -o UserKnownHostsFile=/dev/null -o CheckHostIP=no -o 
StrictHostKeyChecking=no -P  ./reproducer.sh ubuntu@127.0.0.1:~/
  sshpass -p passw0rd ssh -o UserKnownHostsFile=/dev/null -o CheckHostIP=no -o 
StrictHostKeyChecking=no -p  ubuntu@127.0.0.1 sudo apt-get update
  sshpass -p passw0rd ssh -o UserKnownHostsFile=/dev/null -o CheckHostIP=no -o 
StrictHostKeyChecking=no -p  ubuntu@127.0.0.1 sudo apt-get upgrade 
--assume-yes
  sshpass -p passw0rd ssh -o UserKnownHostsFile=/dev/null -o CheckHostIP=no -o 
StrictHostKeyChecking=no -p  ubuntu@127.0.0.1 ./reproducer.sh
  ```

  The issue is not present with the 6.5.0-5 kernel and the issue is
  present regardless of the container launched. I tried the jammy
  container to test this.

  From my test VM

  ```
  ubuntu@cloudimg:~$ uname --all
  Linux cloudimg 6.5.0-7-generic #7-Ubuntu SMP PREEMPT_DYNAMIC Fri Sep 29 
09:14:56 UTC 2023 x86_64 x86_64 x86_64 GNU/Linux
  ubuntu@cloudimg:~$ uname --kernel-release
  6.5.0-7-generic
  ```

  This is a regression in our test that will block 23.10 cloud image
  release next week.

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu-release-notes/+bug/2038567/+subscriptions


-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 2038567] [NEW] Mantic 6.5.0-7 kernel causes regression in LXD container usage

2023-10-05 Thread Philip Roche
Public bug reported:

Following upgrade to 6.5.0-7 kernel in mantic cloud images we are seeing
a regression in our cloud image tests. The test runs the following:

```
lxd init --auto --storage-backend dir
lxc launch ubuntu-daily:mantic mantic
lxc info mantic
lxc exec mantic -- cloud-init status --wait
```

The `lxc exec mantic -- cloud-init status --wait` times out after 240s
and will fail our test as a result.

I have been able to replicate in a local VM

```
wget 
http://cloud-images.ubuntu.com/mantic/20231005/mantic-server-cloudimg-amd64.img 
wget --output-document=launch-qcow2-image-qemu.sh 
https://gist.githubusercontent.com/philroche/14c241c086a5730481e24178b654268f/raw/7af95cd4dfc8e1d0600e6118803d2c866765714e/gistfile1.txt
 
chmod +x launch-qcow2-image-qemu.sh 

./launch-qcow2-image-qemu.sh --password passw0rd --image 
./mantic-server-cloudimg-amd64.img 
cat < "./reproducer.sh"
#!/bin/bash -eux
lxd init --auto --storage-backend dir
lxc launch ubuntu-daily:mantic mantic
lxc info mantic
lxc exec mantic -- cloud-init status --wait
EOF
chmod +x ./reproducer.sh
sshpass -p passw0rd scp -o UserKnownHostsFile=/dev/null -o CheckHostIP=no -o 
StrictHostKeyChecking=no -P  ./reproducer.sh ubuntu@127.0.0.1:~/
sshpass -p passw0rd ssh -o UserKnownHostsFile=/dev/null -o CheckHostIP=no -o 
StrictHostKeyChecking=no -p  ubuntu@127.0.0.1 sudo apt-get update
sshpass -p passw0rd ssh -o UserKnownHostsFile=/dev/null -o CheckHostIP=no -o 
StrictHostKeyChecking=no -p  ubuntu@127.0.0.1 sudo apt-get upgrade 
--assume-yes
sshpass -p passw0rd ssh -o UserKnownHostsFile=/dev/null -o CheckHostIP=no -o 
StrictHostKeyChecking=no -p  ubuntu@127.0.0.1 ./reproducer.sh
```

The issue is not present with the 6.5.0-5 kernel and the issue is
present regardless of the container launched. I tried the jammy
container to test this.

>From my test VM

```
ubuntu@cloudimg:~$ uname --all
Linux cloudimg 6.5.0-7-generic #7-Ubuntu SMP PREEMPT_DYNAMIC Fri Sep 29 
09:14:56 UTC 2023 x86_64 x86_64 x86_64 GNU/Linux
ubuntu@cloudimg:~$ uname --kernel-release
6.5.0-7-generic
```

This is a regression in our test that will block 23.10 cloud image
release next week.

** Affects: ubuntu-release-notes
 Importance: Undecided
 Status: New

** Affects: linux (Ubuntu)
 Importance: Critical
 Status: Incomplete

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/2038567

Title:
  Mantic 6.5.0-7 kernel causes regression in LXD container usage

Status in Release Notes for Ubuntu:
  New
Status in linux package in Ubuntu:
  Incomplete

Bug description:
  Following upgrade to 6.5.0-7 kernel in mantic cloud images we are
  seeing a regression in our cloud image tests. The test runs the
  following:

  ```
  lxd init --auto --storage-backend dir
  lxc launch ubuntu-daily:mantic mantic
  lxc info mantic
  lxc exec mantic -- cloud-init status --wait
  ```

  The `lxc exec mantic -- cloud-init status --wait` times out after 240s
  and will fail our test as a result.

  I have been able to replicate in a local VM

  ```
  wget 
http://cloud-images.ubuntu.com/mantic/20231005/mantic-server-cloudimg-amd64.img 
  wget --output-document=launch-qcow2-image-qemu.sh 
https://gist.githubusercontent.com/philroche/14c241c086a5730481e24178b654268f/raw/7af95cd4dfc8e1d0600e6118803d2c866765714e/gistfile1.txt
 
  chmod +x launch-qcow2-image-qemu.sh 

  ./launch-qcow2-image-qemu.sh --password passw0rd --image 
./mantic-server-cloudimg-amd64.img 
  cat < "./reproducer.sh"
  #!/bin/bash -eux
  lxd init --auto --storage-backend dir
  lxc launch ubuntu-daily:mantic mantic
  lxc info mantic
  lxc exec mantic -- cloud-init status --wait
  EOF
  chmod +x ./reproducer.sh
  sshpass -p passw0rd scp -o UserKnownHostsFile=/dev/null -o CheckHostIP=no -o 
StrictHostKeyChecking=no -P  ./reproducer.sh ubuntu@127.0.0.1:~/
  sshpass -p passw0rd ssh -o UserKnownHostsFile=/dev/null -o CheckHostIP=no -o 
StrictHostKeyChecking=no -p  ubuntu@127.0.0.1 sudo apt-get update
  sshpass -p passw0rd ssh -o UserKnownHostsFile=/dev/null -o CheckHostIP=no -o 
StrictHostKeyChecking=no -p  ubuntu@127.0.0.1 sudo apt-get upgrade 
--assume-yes
  sshpass -p passw0rd ssh -o UserKnownHostsFile=/dev/null -o CheckHostIP=no -o 
StrictHostKeyChecking=no -p  ubuntu@127.0.0.1 ./reproducer.sh
  ```

  The issue is not present with the 6.5.0-5 kernel and the issue is
  present regardless of the container launched. I tried the jammy
  container to test this.

  From my test VM

  ```
  ubuntu@cloudimg:~$ uname --all
  Linux cloudimg 6.5.0-7-generic #7-Ubuntu SMP PREEMPT_DYNAMIC Fri Sep 29 
09:14:56 UTC 2023 x86_64 x86_64 x86_64 GNU/Linux
  ubuntu@cloudimg:~$ uname --kernel-release
  6.5.0-7-generic
  ```

  This is a regression in our test that will block 23.10 cloud image
  release next week.

To manage notifications about this bug go to:

[Kernel-packages] [Bug 2036968] Re: Mantic minimized/minimal cloud images do not receive IP address during provisioning; systemd regression with wait-online

2023-10-04 Thread Philip Roche
I have also successfully verified that -proposed amd64 kernel
`6.5.0-7-generic` results in successful network configuration when
tested using qemu on an amd64 host with older hardware (ThinkPad T460
with 6th gen intel i5 which is the same hardware which we were able to
reproduce the issue on previously). See
https://people.canonical.com/~philroche/20231003-mantic-minimal-
proposed-kernel/amd64/ for cloud-init logs, some debug output and test
image.

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/2036968

Title:
  Mantic minimized/minimal cloud images do not receive IP address during
  provisioning; systemd regression with wait-online

Status in cloud-images:
  New
Status in linux package in Ubuntu:
  Fix Committed
Status in systemd package in Ubuntu:
  Triaged

Bug description:
  Following a recent change from linux-kvm kernel to linux-generic
  kernel in the mantic minimized images, there is a reproducable bug
  where a guest VM does not have an IP address assigned as part of
  cloud-init provisioning.

  This is easiest to reproduce when emulating arm64 on amd64 host. The
  bug is a race condition, so there could exist fast enough
  virtualisation on fast enough hardware where this bug is not present
  but in all my testing I have been able to reproduce.

  The latest mantic minimized images from http://cloud-
  images.ubuntu.com/minimal/daily/mantic/ have force initrdless boot and
  no initrd to fallback to.

  This but is not present in the non minimized/base images @
  http://cloud-images.ubuntu.com/mantic/ as these boot with initrd with
  the required drivers present for virtio-net.

  Reproducer

  ```
  wget -O "launch-qcow2-image-qemu-arm64.sh" 
https://people.canonical.com/~philroche/20230921-cloud-images-mantic-fail-to-provision/launch-qcow2-image-qemu-arm64.sh

  chmod +x ./launch-qcow2-image-qemu-arm64.sh
  wget 
https://people.canonical.com/~philroche/20230921-cloud-images-mantic-fail-to-provision/livecd.ubuntu-cpc.img
  ./launch-qcow2-image-qemu-arm64.sh --password passw0rd --image 
./livecd.ubuntu-cpc.img
  ```

  You will then be able to log in with user `ubuntu` and password
  `passw0rd`.

  You can run `ip a` and see that there is a network interface present
  (separate to `lo`) but no IP address has been assigned.

  ```
  ubuntu@cloudimg:~$ ip a
  1: lo:  mtu 65536 qdisc noqueue state UNKNOWN group 
default qlen 1000
  link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
  inet 127.0.0.1/8 scope host lo
     valid_lft forever preferred_lft forever
  inet6 ::1/128 scope host
     valid_lft forever preferred_lft forever
  2: enp0s1:  mtu 1500 qdisc noop state DOWN group default 
qlen 1000
  link/ether 52:54:00:12:34:56 brd ff:ff:ff:ff:ff:ff

  ```

  This is because when cloud-init is trying to configure network
  interfaces it doesn't find any so it doesn't configure any. But by the
  time boot is complete the network interface is present but cloud-init
  provisioning has already completed.

  You can verify this by running `sudo cloud-init clean && sudo cloud-
  init init`

  You can then see a successfully configured network interface

  ```
  ubuntu@cloudimg:~$ ip a
  1: lo:  mtu 65536 qdisc noqueue state UNKNOWN group 
default qlen 1000
  link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
  inet 127.0.0.1/8 scope host lo
     valid_lft forever preferred_lft forever
  inet6 ::1/128 scope host
     valid_lft forever preferred_lft forever
  2: enp0s1:  mtu 1500 qdisc pfifo_fast state 
UP group default qlen 1000
  link/ether 52:54:00:12:34:56 brd ff:ff:ff:ff:ff:ff
  inet 10.0.2.15/24 metric 100 brd 10.0.2.255 scope global dynamic enp0s1
     valid_lft 86391sec preferred_lft 86391sec
  inet6 fec0::5054:ff:fe12:3456/64 scope site dynamic mngtmpaddr 
noprefixroute
     valid_lft 86393sec preferred_lft 14393sec
  inet6 fe80::5054:ff:fe12:3456/64 scope link
     valid_lft forever preferred_lft forever

  ```

  The bug is also reproducible with amd64 guest on adm64 host on
  older/slower hardware.

  The suggested fixes while debugging this issue are:

  * to include `virtio-net` as a built-in in the mantic generic kernel
  * understand what needs to change in cloud-init so that it can react to late 
additions of network interfaces

  I will file a separate bug against cloud-init to address the race
  condition on emulated guest/older hardware.

To manage notifications about this bug go to:
https://bugs.launchpad.net/cloud-images/+bug/2036968/+subscriptions


-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 2036968] Re: Mantic minimized/minimal cloud images do not receive IP address during provisioning; systemd regression with wait-online

2023-10-04 Thread Philip Roche
@xnox I have successfully verified that -proposed arm64 kernel
`6.5.0-7-generic` results in successful network configuration when
tested using qemu on an amd64 host. See
https://people.canonical.com/~philroche/20231003-mantic-minimal-
proposed-kernel/ for cloud-init logs, some debug output and test image.

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/2036968

Title:
  Mantic minimized/minimal cloud images do not receive IP address during
  provisioning; systemd regression with wait-online

Status in cloud-images:
  New
Status in linux package in Ubuntu:
  Fix Committed
Status in systemd package in Ubuntu:
  Triaged

Bug description:
  Following a recent change from linux-kvm kernel to linux-generic
  kernel in the mantic minimized images, there is a reproducable bug
  where a guest VM does not have an IP address assigned as part of
  cloud-init provisioning.

  This is easiest to reproduce when emulating arm64 on amd64 host. The
  bug is a race condition, so there could exist fast enough
  virtualisation on fast enough hardware where this bug is not present
  but in all my testing I have been able to reproduce.

  The latest mantic minimized images from http://cloud-
  images.ubuntu.com/minimal/daily/mantic/ have force initrdless boot and
  no initrd to fallback to.

  This but is not present in the non minimized/base images @
  http://cloud-images.ubuntu.com/mantic/ as these boot with initrd with
  the required drivers present for virtio-net.

  Reproducer

  ```
  wget -O "launch-qcow2-image-qemu-arm64.sh" 
https://people.canonical.com/~philroche/20230921-cloud-images-mantic-fail-to-provision/launch-qcow2-image-qemu-arm64.sh

  chmod +x ./launch-qcow2-image-qemu-arm64.sh
  wget 
https://people.canonical.com/~philroche/20230921-cloud-images-mantic-fail-to-provision/livecd.ubuntu-cpc.img
  ./launch-qcow2-image-qemu-arm64.sh --password passw0rd --image 
./livecd.ubuntu-cpc.img
  ```

  You will then be able to log in with user `ubuntu` and password
  `passw0rd`.

  You can run `ip a` and see that there is a network interface present
  (separate to `lo`) but no IP address has been assigned.

  ```
  ubuntu@cloudimg:~$ ip a
  1: lo:  mtu 65536 qdisc noqueue state UNKNOWN group 
default qlen 1000
  link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
  inet 127.0.0.1/8 scope host lo
     valid_lft forever preferred_lft forever
  inet6 ::1/128 scope host
     valid_lft forever preferred_lft forever
  2: enp0s1:  mtu 1500 qdisc noop state DOWN group default 
qlen 1000
  link/ether 52:54:00:12:34:56 brd ff:ff:ff:ff:ff:ff

  ```

  This is because when cloud-init is trying to configure network
  interfaces it doesn't find any so it doesn't configure any. But by the
  time boot is complete the network interface is present but cloud-init
  provisioning has already completed.

  You can verify this by running `sudo cloud-init clean && sudo cloud-
  init init`

  You can then see a successfully configured network interface

  ```
  ubuntu@cloudimg:~$ ip a
  1: lo:  mtu 65536 qdisc noqueue state UNKNOWN group 
default qlen 1000
  link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
  inet 127.0.0.1/8 scope host lo
     valid_lft forever preferred_lft forever
  inet6 ::1/128 scope host
     valid_lft forever preferred_lft forever
  2: enp0s1:  mtu 1500 qdisc pfifo_fast state 
UP group default qlen 1000
  link/ether 52:54:00:12:34:56 brd ff:ff:ff:ff:ff:ff
  inet 10.0.2.15/24 metric 100 brd 10.0.2.255 scope global dynamic enp0s1
     valid_lft 86391sec preferred_lft 86391sec
  inet6 fec0::5054:ff:fe12:3456/64 scope site dynamic mngtmpaddr 
noprefixroute
     valid_lft 86393sec preferred_lft 14393sec
  inet6 fe80::5054:ff:fe12:3456/64 scope link
     valid_lft forever preferred_lft forever

  ```

  The bug is also reproducible with amd64 guest on adm64 host on
  older/slower hardware.

  The suggested fixes while debugging this issue are:

  * to include `virtio-net` as a built-in in the mantic generic kernel
  * understand what needs to change in cloud-init so that it can react to late 
additions of network interfaces

  I will file a separate bug against cloud-init to address the race
  condition on emulated guest/older hardware.

To manage notifications about this bug go to:
https://bugs.launchpad.net/cloud-images/+bug/2036968/+subscriptions


-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 2037398] Re: kexec enable to load/kdump zstd compressed zimg

2023-09-26 Thread Philip Roche
I have confirmed that this issue with not being able to capture kernel
dump with a mantic arm64 kernel is not new. using the arm64 6.5 kernel
(6.5.0-5) in the release pocket I captured the following during test.

```
ubuntu@cloudimg:~$ sudo kdump-config show
DUMP_MODE:  kdump
USE_KDUMP:  1
KDUMP_COREDIR:  /var/crash
crashkernel addr: 0xde00
   /var/lib/kdump/vmlinuz: symbolic link to /boot/vmlinuz-6.5.0-5-generic
kdump initrd: 
   /var/lib/kdump/initrd.img: symbolic link to 
/var/lib/kdump/initrd.img-6.5.0-5-generic
current state:Not ready to kdump

kexec command:
  no kexec command recorded
ubuntu@cloudimg:~$ uname --all
Linux cloudimg 6.5.0-5-generic #5-Ubuntu SMP PREEMPT_DYNAMIC Wed Sep  6 
15:36:23 UTC 2023 aarch64 aarch64 aarch64 GNU/Linux
ubuntu@cloudimg:~$ sudo cat /proc/cmdline
BOOT_IMAGE=/vmlinuz-6.5.0-5-generic 
root=UUID=e7604ab2-200c-4f34-ab11-47e78ac4b8bd ro console=tty1 console=ttyS0 
crashkernel=2G-4G:320M,4G-32G:512M,32G-64G:1024M,64G-128G:2048M,128G-:4096M
ubuntu@cloudimg:~$ sudo dmesg | grep -i crash
[0.00] crashkernel reserved: 0xde00 - 0xfe00 
(512 MB)
[0.00] Kernel command line: BOOT_IMAGE=/vmlinuz-6.5.0-5-generic 
root=UUID=e7604ab2-200c-4f34-ab11-47e78ac4b8bd ro console=tty1 console=ttyS0 
crashkernel=2G-4G:320M,4G-32G:512M,32G-64G:1024M,64G-128G:2048M,128G-:4096M
[   87.054802] pstore: Using crash dump compression: deflate
ubuntu@cloudimg:~$ sudo cat /proc/sys/kernel/sysrq
176
ubuntu@cloudimg:~$ sudo service kdump-tools status
● kdump-tools.service - Kernel crash dump capture service
 Loaded: loaded (/lib/systemd/system/kdump-tools.service; enabled; preset: 
enabled)
 Active: active (exited) since Tue 2023-09-26 13:44:22 UTC; 7min ago
Process: 514 ExecStart=/etc/init.d/kdump-tools start (code=exited, 
status=0/SUCCESS)
   Main PID: 514 (code=exited, status=0/SUCCESS)
CPU: 4min 46.461s

Sep 26 13:38:21 cloudimg systemd[1]: Starting kdump-tools.service - Kernel 
crash dump capture service...
Sep 26 13:38:32 cloudimg kdump-tools[514]: Starting kdump-tools:
Sep 26 13:38:32 cloudimg kdump-tools[535]:  * Creating symlink 
/var/lib/kdump/vmlinuz
Sep 26 13:38:41 cloudimg kdump-tools[577]: kdump-tools: Generating 
/var/lib/kdump/initrd.img-6.5.0-5-generic
Sep 26 13:44:11 cloudimg kdump-tools[535]:  * Creating symlink 
/var/lib/kdump/initrd.img
Sep 26 13:44:22 cloudimg kdump-tools[535]:  * failed to load kdump kernel
Sep 26 13:44:22 cloudimg kdump-tools[5536]: failed to load kdump kernel
Sep 26 13:44:22 cloudimg systemd[1]: Finished kdump-tools.service - Kernel 
crash dump capture service.
ubuntu@cloudimg:~$ sudo ls -al /var/lib/kdump/vmlinuz 
lrwxrwxrwx 1 root root 29 Sep 26 13:38 /var/lib/kdump/vmlinuz -> 
/boot/vmlinuz-6.5.0-5-generic
ubuntu@cloudimg:~$ sudo file /boot/vmlinuz-6.5.0-5-generic
/boot/vmlinuz-6.5.0-5-generic: gzip compressed data, was 
"vmlinuz-6.5.0-5-generic.efi.signed", last modified: Thu Sep  7 10:43:26 2023, 
max compression, from Unix, original size modulo 2^32 54989192
```

.. but the `Cannot determine the file type of /var/lib/kdump/vmlinuz` in
the `sudo service kdump-tools status` as @xnox has also reproduced with
`kexec` is new.

I have also confirmed that capturing kernel crash dump with the amd64
kernel 6.5.0-5 & 6.5.0-6 works as expected.

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to kexec-tools in Ubuntu.
https://bugs.launchpad.net/bugs/2037398

Title:
  kexec enable to load/kdump zstd compressed zimg

Status in kdump-tools package in Ubuntu:
  Invalid
Status in kexec-tools package in Ubuntu:
  Triaged
Status in linux package in Ubuntu:
  Triaged

Bug description:
  While testing the 6.5.0-6-generic proposed arm64 generic kernel I
  encountered issues being able to use kdump.

  After enabling -proposed and installing the -proposed 6.5.0-6 kernel
  and rebooting I encountered the following:

  `kdump-config show` shows `current state:Not ready to kdump` and
  looking at the status of the kdump-tools services I see `Cannot
  determine the file type of /var/lib/kdump/vmlinuz`

  Full output:

  ```
  ubuntu@cloudimg:~$ sudo kdump-config show
  DUMP_MODE:kdump
  USE_KDUMP:1
  KDUMP_COREDIR:/var/crash
  crashkernel addr: 0xde00
 /var/lib/kdump/vmlinuz: symbolic link to /boot/vmlinuz-6.5.0-6-generic
  kdump initrd: 
 /var/lib/kdump/initrd.img: symbolic link to 
/var/lib/kdump/initrd.img-6.5.0-6-generic
  current state:Not ready to kdump

  kexec command:
no kexec command recorded
  ubuntu@cloudimg:~$ sudo service kdump-tools status
  ● kdump-tools.service - Kernel crash dump capture service
   Loaded: loaded (/lib/systemd/system/kdump-tools.service; enabled; 
preset: enabled)
   Active: active (exited) since Tue 2023-09-26 09:21:44 UTC; 5min ago
  Process: 515 ExecStart=/etc/init.d/kdump-tools start (code=exited, 

[Kernel-packages] [Bug 2037398] [NEW] Unable to capture kernel crash dump using arm64 mantic 6.5.0-6-generic -proposed kernel

2023-09-26 Thread Philip Roche
Public bug reported:

While testing the 6.5.0-6-generic proposed arm64 generic kernel I
encountered issues being able to use kdump.

After enabling -proposed and installing the -proposed 6.5.0-6 kernel and
rebooting I encountered the following:

`kdump-config show` shows `current state:Not ready to kdump` and
looking at the status of the kdump-tools services I see `Cannot
determine the file type of /var/lib/kdump/vmlinuz`

Full output:

```
ubuntu@cloudimg:~$ sudo kdump-config show
DUMP_MODE:  kdump
USE_KDUMP:  1
KDUMP_COREDIR:  /var/crash
crashkernel addr: 0xde00
   /var/lib/kdump/vmlinuz: symbolic link to /boot/vmlinuz-6.5.0-6-generic
kdump initrd: 
   /var/lib/kdump/initrd.img: symbolic link to 
/var/lib/kdump/initrd.img-6.5.0-6-generic
current state:Not ready to kdump

kexec command:
  no kexec command recorded
ubuntu@cloudimg:~$ sudo service kdump-tools status
● kdump-tools.service - Kernel crash dump capture service
 Loaded: loaded (/lib/systemd/system/kdump-tools.service; enabled; preset: 
enabled)
 Active: active (exited) since Tue 2023-09-26 09:21:44 UTC; 5min ago
Process: 515 ExecStart=/etc/init.d/kdump-tools start (code=exited, 
status=0/SUCCESS)
   Main PID: 515 (code=exited, status=0/SUCCESS)
CPU: 4min 21.329s

Sep 26 09:16:14 cloudimg systemd[1]: Starting kdump-tools.service - Kernel 
crash dump capture service...
Sep 26 09:16:24 cloudimg kdump-tools[515]: Starting kdump-tools:
Sep 26 09:16:24 cloudimg kdump-tools[537]:  * Creating symlink 
/var/lib/kdump/vmlinuz
Sep 26 09:16:32 cloudimg kdump-tools[580]: kdump-tools: Generating 
/var/lib/kdump/initrd.img-6.5.0-6-generic
Sep 26 09:21:42 cloudimg kdump-tools[537]:  * Creating symlink 
/var/lib/kdump/initrd.img
Sep 26 09:21:43 cloudimg kdump-tools[5538]: Cannot determine the file type of 
/var/lib/kdump/vmlinuz
Sep 26 09:21:44 cloudimg kdump-tools[537]:  * failed to load kdump kernel
Sep 26 09:21:44 cloudimg kdump-tools[5539]: failed to load kdump kernel
Sep 26 09:21:44 cloudimg systemd[1]: Finished kdump-tools.service - Kernel 
crash dump capture service.

ubuntu@cloudimg:~$ ls -al /var/lib/kdump/vmlinuz
lrwxrwxrwx 1 root root 29 Sep 26 09:16 /var/lib/kdump/vmlinuz -> 
/boot/vmlinuz-6.5.0-6-generic
ubuntu@cloudimg:~$ file /var/lib/kdump/vmlinuz
/var/lib/kdump/vmlinuz: symbolic link to /boot/vmlinuz-6.5.0-6-generic
ubuntu@cloudimg:~$ sudo file /boot/vmlinuz-6.5.0-6-generic
/boot/vmlinuz-6.5.0-6-generic: PE32+ executable (EFI application) Aarch64 
(stripped to external PDB), for MS Windows, 2 sections
```

The reboot with 6.5.0-6 was successful and the reboot after linux-crashdump 
install was successful too. 
I used https://ubuntu.com/server/docs/kernel-crash-dump guide for installing 
linux-crashdump and attempting to trigger a dump.

I used arm64 qcow cloud image from http://cloud-
images.ubuntu.com/mantic/20230925/mantic-server-cloudimg-arm64.img to
test the above emulated on amd64.

** Affects: linux (Ubuntu)
 Importance: Undecided
 Status: New

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/2037398

Title:
  Unable to capture kernel crash dump using arm64 mantic 6.5.0-6-generic
  -proposed kernel

Status in linux package in Ubuntu:
  New

Bug description:
  While testing the 6.5.0-6-generic proposed arm64 generic kernel I
  encountered issues being able to use kdump.

  After enabling -proposed and installing the -proposed 6.5.0-6 kernel
  and rebooting I encountered the following:

  `kdump-config show` shows `current state:Not ready to kdump` and
  looking at the status of the kdump-tools services I see `Cannot
  determine the file type of /var/lib/kdump/vmlinuz`

  Full output:

  ```
  ubuntu@cloudimg:~$ sudo kdump-config show
  DUMP_MODE:kdump
  USE_KDUMP:1
  KDUMP_COREDIR:/var/crash
  crashkernel addr: 0xde00
 /var/lib/kdump/vmlinuz: symbolic link to /boot/vmlinuz-6.5.0-6-generic
  kdump initrd: 
 /var/lib/kdump/initrd.img: symbolic link to 
/var/lib/kdump/initrd.img-6.5.0-6-generic
  current state:Not ready to kdump

  kexec command:
no kexec command recorded
  ubuntu@cloudimg:~$ sudo service kdump-tools status
  ● kdump-tools.service - Kernel crash dump capture service
   Loaded: loaded (/lib/systemd/system/kdump-tools.service; enabled; 
preset: enabled)
   Active: active (exited) since Tue 2023-09-26 09:21:44 UTC; 5min ago
  Process: 515 ExecStart=/etc/init.d/kdump-tools start (code=exited, 
status=0/SUCCESS)
 Main PID: 515 (code=exited, status=0/SUCCESS)
  CPU: 4min 21.329s

  Sep 26 09:16:14 cloudimg systemd[1]: Starting kdump-tools.service - Kernel 
crash dump capture service...
  Sep 26 09:16:24 cloudimg kdump-tools[515]: Starting kdump-tools:
  Sep 26 09:16:24 cloudimg kdump-tools[537]:  * Creating symlink 
/var/lib/kdump/vmlinuz
  Sep 26 09:16:32 cloudimg 

[Kernel-packages] [Bug 2036968] Re: Mantic minimized/minimal cloud images do not receive IP address during provisioning

2023-09-21 Thread Philip Roche
cloud-init bug filed @ https://github.com/canonical/cloud-
init/issues/4451

** Bug watch added: github.com/canonical/cloud-init/issues #4451
   https://github.com/canonical/cloud-init/issues/4451

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/2036968

Title:
  Mantic minimized/minimal cloud images do not receive IP address during
  provisioning

Status in cloud-images:
  New
Status in linux package in Ubuntu:
  Incomplete

Bug description:
  Following a recent change from linux-kvm kernel to linux-generic
  kernel in the mantic minimized images, there is a reproducable bug
  where a guest VM does not have an IP address assigned as part of
  cloud-init provisioning.

  This is easiest to reproduce when emulating arm64 on amd64 host. The
  bug is a race condition, so there could exist fast enough
  virtualisation on fast enough hardware where this bug is not present
  but in all my testing I have been able to reproduce.

  The latest mantic minimized images from http://cloud-
  images.ubuntu.com/minimal/daily/mantic/ have force initrdless boot and
  no initrd to fallback to.

  This but is not present in the non minimized/base images @
  http://cloud-images.ubuntu.com/mantic/ as these boot with initrd with
  the required drivers present for virtio-net.

  Reproducer

  ```
  wget -O "launch-qcow2-image-qemu-arm64.sh" 
https://people.canonical.com/~philroche/20230921-cloud-images-mantic-fail-to-provision/launch-qcow2-image-qemu-arm64.sh

  chmod +x ./launch-qcow2-image-qemu-arm64.sh
  wget 
https://people.canonical.com/~philroche/20230921-cloud-images-mantic-fail-to-provision/livecd.ubuntu-cpc.img
  ./launch-qcow2-image-qemu-arm64.sh --password passw0rd --image 
./livecd.ubuntu-cpc.img
  ```

  You will then be able to log in with user `ubuntu` and password
  `passw0rd`.

  You can run `ip a` and see that there is a network interface present
  (separate to `lo`) but no IP address has been assigned.

  ```
  ubuntu@cloudimg:~$ ip a
  1: lo:  mtu 65536 qdisc noqueue state UNKNOWN group 
default qlen 1000
  link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
  inet 127.0.0.1/8 scope host lo
     valid_lft forever preferred_lft forever
  inet6 ::1/128 scope host
     valid_lft forever preferred_lft forever
  2: enp0s1:  mtu 1500 qdisc noop state DOWN group default 
qlen 1000
  link/ether 52:54:00:12:34:56 brd ff:ff:ff:ff:ff:ff

  ```

  This is because when cloud-init is trying to configure network
  interfaces it doesn't find any so it doesn't configure any. But by the
  time boot is complete the network interface is present but cloud-init
  provisioning has already completed.

  You can verify this by running `sudo cloud-init clean && sudo cloud-
  init init`

  You can then see a successfully configured network interface

  ```
  ubuntu@cloudimg:~$ ip a
  1: lo:  mtu 65536 qdisc noqueue state UNKNOWN group 
default qlen 1000
  link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
  inet 127.0.0.1/8 scope host lo
     valid_lft forever preferred_lft forever
  inet6 ::1/128 scope host
     valid_lft forever preferred_lft forever
  2: enp0s1:  mtu 1500 qdisc pfifo_fast state 
UP group default qlen 1000
  link/ether 52:54:00:12:34:56 brd ff:ff:ff:ff:ff:ff
  inet 10.0.2.15/24 metric 100 brd 10.0.2.255 scope global dynamic enp0s1
     valid_lft 86391sec preferred_lft 86391sec
  inet6 fec0::5054:ff:fe12:3456/64 scope site dynamic mngtmpaddr 
noprefixroute
     valid_lft 86393sec preferred_lft 14393sec
  inet6 fe80::5054:ff:fe12:3456/64 scope link
     valid_lft forever preferred_lft forever

  ```

  The bug is also reproducible with amd64 guest on adm64 host on
  older/slower hardware.

  The suggested fixes while debugging this issue are:

  * to include `virtio-net` as a built-in in the mantic generic kernel
  * understand what needs to change in cloud-init so that it can react to late 
additions of network interfaces

  I will file a separate bug against cloud-init to address the race
  condition on emulated guest/older hardware.

To manage notifications about this bug go to:
https://bugs.launchpad.net/cloud-images/+bug/2036968/+subscriptions


-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 2036968] Re: Mantic minimized/minimal cloud images do not receive IP address during provisioning

2023-09-21 Thread Philip Roche
** Description changed:

  Following a recent change from linux-kvm kernel to linux-generic kernel
- in the mantic minimized images there is a reproducable bug where a guest
- VM does not have an IP address assigned as part of cloud-init
+ in the mantic minimized images, there is a reproducable bug where a
+ guest VM does not have an IP address assigned as part of cloud-init
  provisioning.
  
- This is easiest to reproduce when emulating arm64 on adm64 host. The bug
- is a race condition so there could exist fast enough virtualisation on
+ This is easiest to reproduce when emulating arm64 on amd64 host. The bug
+ is a race condition, so there could exist fast enough virtualisation on
  fast enough hardware where this bug is not present but in all my testing
  I have been able to reproduce.
  
  The latest mantic minimized images from http://cloud-
  images.ubuntu.com/minimal/daily/mantic/ have force initrdless boot and
  no initrd to fallback to.
  
- This bug is not present in the non minimized/base images @ http://cloud-
+ This but is not present in the non minimized/base images @ http://cloud-
  images.ubuntu.com/mantic/ as these boot with initrd with the required
  drivers present for virtio-net.
  
  Reproducer
  
  ```
  wget -O "launch-qcow2-image-qemu-arm64.sh" 
https://people.canonical.com/~philroche/20230921-cloud-images-mantic-fail-to-provision/launch-qcow2-image-qemu-arm64.sh
  
  chmod +x ./launch-qcow2-image-qemu-arm64.sh
  wget 
https://people.canonical.com/~philroche/20230921-cloud-images-mantic-fail-to-provision/livecd.ubuntu-cpc.img
  ./launch-qcow2-image-qemu-arm64.sh --password passw0rd --image 
./livecd.ubuntu-cpc.img
  ```
  
  You will then be able to log in with user `ubuntu` and password
  `passw0rd`.
  
  You can run `ip a` and see that there is a network interface present
  (separate to `lo`) but no IP address has been assigned.
  
  ```
  ubuntu@cloudimg:~$ ip a
  1: lo:  mtu 65536 qdisc noqueue state UNKNOWN group 
default qlen 1000
  link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
  inet 127.0.0.1/8 scope host lo
     valid_lft forever preferred_lft forever
  inet6 ::1/128 scope host
     valid_lft forever preferred_lft forever
  2: enp0s1:  mtu 1500 qdisc noop state DOWN group default 
qlen 1000
  link/ether 52:54:00:12:34:56 brd ff:ff:ff:ff:ff:ff
  
  ```
  
- This is because when cloud-init-local.service is trying to configure
- network interfaces it doesn't find any so it doesn't configure any. But
- by the time boot is complete the network interface is present but cloud-
- init provisioning has already completed.
+ This is because when cloud-init is trying to configure network
+ interfaces it doesn't find any so it doesn't configure any. But by the
+ time boot is complete the network interface is present but cloud-init
+ provisioning has already completed.
  
  You can verify this by running `sudo cloud-init clean && sudo cloud-init
  init`
  
  You can then see a successfully configured network interface
  
  ```
  ubuntu@cloudimg:~$ ip a
  1: lo:  mtu 65536 qdisc noqueue state UNKNOWN group 
default qlen 1000
  link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
  inet 127.0.0.1/8 scope host lo
     valid_lft forever preferred_lft forever
  inet6 ::1/128 scope host
     valid_lft forever preferred_lft forever
  2: enp0s1:  mtu 1500 qdisc pfifo_fast state 
UP group default qlen 1000
  link/ether 52:54:00:12:34:56 brd ff:ff:ff:ff:ff:ff
  inet 10.0.2.15/24 metric 100 brd 10.0.2.255 scope global dynamic enp0s1
     valid_lft 86391sec preferred_lft 86391sec
  inet6 fec0::5054:ff:fe12:3456/64 scope site dynamic mngtmpaddr 
noprefixroute
     valid_lft 86393sec preferred_lft 14393sec
  inet6 fe80::5054:ff:fe12:3456/64 scope link
     valid_lft forever preferred_lft forever
  
  ```
  
- The bug is also reproducable with amd64 guest on adm64 host on
+ The bug is also reproducible with amd64 guest on adm64 host on
  older/slower hardware.
  
  The suggested fixes while debugging this issue are:
  
  * to include `virtio-net` as a built-in in the mantic generic kernel
  * understand what needs to change in cloud-init so that it can react to late 
additions of network interfaces
  
  I will file a separate bug against cloud-init to address the race
  condition on emulated guest/older hardware.

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/2036968

Title:
  Mantic minimized/minimal cloud images do not receive IP address during
  provisioning

Status in cloud-images:
  New
Status in linux package in Ubuntu:
  Incomplete

Bug description:
  Following a recent change from linux-kvm kernel to linux-generic
  kernel in the mantic minimized images, there is a reproducable bug
  where a guest VM does not have an IP address assigned as part of
  cloud-init provisioning.

  This is easiest to 

[Kernel-packages] [Bug 2032933] Re: Mantic (23.10) minimal images increase in memory consumption, port usage and processes running

2023-09-21 Thread Philip Roche
There is a related bug @
https://bugs.launchpad.net/ubuntu/+source/linux/+bug/2036968 which might
have affected boot speed.

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/2032933

Title:
  Mantic (23.10) minimal images increase in memory consumption, port
  usage and processes running

Status in cloud-images:
  Confirmed
Status in linux package in Ubuntu:
  Confirmed

Bug description:
  The Mantic (Ubuntu 23.10) download/qcow2 images available @ 
https://cloud-images.ubuntu.com/minimal/
  are undergoing some big changes prior to 23.10 release in October.

  This is a devel release so this is the perfect time to be making these
  changes but we are noticing some changes that were not expected.

  This bug is to track the unexpected changes and discuss/resolve these.

  The changes that have been made to mantic minimal:

  * Move to the linux-generic kernel from the linux-kvm kernel
    * This also involved removal of the virtio-blk driver, which is the default 
for QEMU and OpenStack, but this is being restored in an upcoming 6.5 mantic 
kernel and is being trakced @ 
https://bugs.launchpad.net/ubuntu/+source/linux/+bug/2030745
  * Move to using minimal-cloud seed - see 
https://ubuntu-archive-team.ubuntu.com/seeds/ubuntu.mantic/cloud-minimal
  * No longer installing Recommends packages
* This is during image build only and will not affect any subsequent 
package installs
  * No initramfs fallback for boot - only initramfsless boot

  The latest mantic minimal images are available @ http://cloud-
  images.ubuntu.com/minimal/daily/mantic/ and are also available in the
  public clouds.

  A package name manifest diff can be seen @
  https://pastebin.ubuntu.com/p/rRd6STnNmK/

  We have had reports of higher memory usage on an idle system, higher
  number of ports open on an idle system and higher number of process
  running on a idle system.

  To help with debugging I have built and uploaded the following images
  and package manifests to
  https://people.canonical.com/~philroche/20230824-manticl-minimal-
  LP2032933/

  * 
20230618-before-kernel-change-before-seed-change-mantic-minimal-cloudimg-amd64
    * Before kernel change and before seed change
  * 
20230824-after-kernel-change-before-seed-change-mantic-minimal-cloudimg-amd64
    * After kernel change and before seed change
  * 
20230821.1-after-kernel-change-after-seed-change-mantic-minimal-cloudimg-amd64
    * After kernel change and after seed change

To manage notifications about this bug go to:
https://bugs.launchpad.net/cloud-images/+bug/2032933/+subscriptions


-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 2036968] Re: Mantic minimized/minimal cloud images do not receive IP address during provisioning

2023-09-21 Thread Philip Roche
** Also affects: linux (Ubuntu)
   Importance: Undecided
   Status: New

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/2036968

Title:
  Mantic minimized/minimal cloud images do not receive IP address during
  provisioning

Status in cloud-images:
  New
Status in linux package in Ubuntu:
  New

Bug description:
  Following a recent change from linux-kvm kernel to linux-generic
  kernel in the mantic minimized images there is a reproducable bug
  where a guest VM does not have an IP address assigned as part of
  cloud-init provisioning.

  This is easiest to reproduce when emulating arm64 on adm64 host. The
  bug is a race condition so there could exist fast enough
  virtualisation on fast enough hardware where this bug is not present
  but in all my testing I have been able to reproduce.

  The latest mantic minimized images from http://cloud-
  images.ubuntu.com/minimal/daily/mantic/ have force initrdless boot and
  no initrd to fallback to.

  This bug is not present in the non minimized/base images @
  http://cloud-images.ubuntu.com/mantic/ as these boot with initrd with
  the required drivers present for virtio-net.

  Reproducer

  ```
  wget -O "launch-qcow2-image-qemu-arm64.sh" 
https://people.canonical.com/~philroche/20230921-cloud-images-mantic-fail-to-provision/launch-qcow2-image-qemu-arm64.sh

  chmod +x ./launch-qcow2-image-qemu-arm64.sh
  wget 
https://people.canonical.com/~philroche/20230921-cloud-images-mantic-fail-to-provision/livecd.ubuntu-cpc.img
  ./launch-qcow2-image-qemu-arm64.sh --password passw0rd --image 
./livecd.ubuntu-cpc.img
  ```

  You will then be able to log in with user `ubuntu` and password
  `passw0rd`.

  You can run `ip a` and see that there is a network interface present
  (separate to `lo`) but no IP address has been assigned.

  ```
  ubuntu@cloudimg:~$ ip a
  1: lo:  mtu 65536 qdisc noqueue state UNKNOWN group 
default qlen 1000
  link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
  inet 127.0.0.1/8 scope host lo
     valid_lft forever preferred_lft forever
  inet6 ::1/128 scope host
     valid_lft forever preferred_lft forever
  2: enp0s1:  mtu 1500 qdisc noop state DOWN group default 
qlen 1000
  link/ether 52:54:00:12:34:56 brd ff:ff:ff:ff:ff:ff

  ```

  This is because when cloud-init is trying to configure network
  interfaces it doesn't find any so it doesn't configure any. But by the
  time boot is complete the network interface is present but cloud-init
  provisioning has already completed.

  You can verify this by running `sudo cloud-init clean && sudo cloud-
  init init`

  You can then see a successfully configured network interface

  ```
  ubuntu@cloudimg:~$ ip a
  1: lo:  mtu 65536 qdisc noqueue state UNKNOWN group 
default qlen 1000
  link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
  inet 127.0.0.1/8 scope host lo
     valid_lft forever preferred_lft forever
  inet6 ::1/128 scope host
     valid_lft forever preferred_lft forever
  2: enp0s1:  mtu 1500 qdisc pfifo_fast state 
UP group default qlen 1000
  link/ether 52:54:00:12:34:56 brd ff:ff:ff:ff:ff:ff
  inet 10.0.2.15/24 metric 100 brd 10.0.2.255 scope global dynamic enp0s1
     valid_lft 86391sec preferred_lft 86391sec
  inet6 fec0::5054:ff:fe12:3456/64 scope site dynamic mngtmpaddr 
noprefixroute
     valid_lft 86393sec preferred_lft 14393sec
  inet6 fe80::5054:ff:fe12:3456/64 scope link
     valid_lft forever preferred_lft forever

  ```

  The bug is also reproducable with amd64 guest on adm64 host on
  older/slower hardware.

  The suggested fixes while debugging this issue are:

  * to include `virtio-net` as a built-in in the mantic generic kernel
  * understand what needs to change in cloud-init so that it can react to late 
additions of network interfaces

  I will file a separate bug against cloud-init to address the race
  condition on emulated guest/older hardware.

To manage notifications about this bug go to:
https://bugs.launchpad.net/cloud-images/+bug/2036968/+subscriptions


-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 2032933] Re: Mantic (23.10) minimal images increase in memory consumption, port usage and processes running

2023-09-21 Thread Philip Roche
@paelzer agreed. Good plan.

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/2032933

Title:
  Mantic (23.10) minimal images increase in memory consumption, port
  usage and processes running

Status in cloud-images:
  Confirmed
Status in linux package in Ubuntu:
  Confirmed

Bug description:
  The Mantic (Ubuntu 23.10) download/qcow2 images available @ 
https://cloud-images.ubuntu.com/minimal/
  are undergoing some big changes prior to 23.10 release in October.

  This is a devel release so this is the perfect time to be making these
  changes but we are noticing some changes that were not expected.

  This bug is to track the unexpected changes and discuss/resolve these.

  The changes that have been made to mantic minimal:

  * Move to the linux-generic kernel from the linux-kvm kernel
    * This also involved removal of the virtio-blk driver, which is the default 
for QEMU and OpenStack, but this is being restored in an upcoming 6.5 mantic 
kernel and is being trakced @ 
https://bugs.launchpad.net/ubuntu/+source/linux/+bug/2030745
  * Move to using minimal-cloud seed - see 
https://ubuntu-archive-team.ubuntu.com/seeds/ubuntu.mantic/cloud-minimal
  * No longer installing Recommends packages
* This is during image build only and will not affect any subsequent 
package installs
  * No initramfs fallback for boot - only initramfsless boot

  The latest mantic minimal images are available @ http://cloud-
  images.ubuntu.com/minimal/daily/mantic/ and are also available in the
  public clouds.

  A package name manifest diff can be seen @
  https://pastebin.ubuntu.com/p/rRd6STnNmK/

  We have had reports of higher memory usage on an idle system, higher
  number of ports open on an idle system and higher number of process
  running on a idle system.

  To help with debugging I have built and uploaded the following images
  and package manifests to
  https://people.canonical.com/~philroche/20230824-manticl-minimal-
  LP2032933/

  * 
20230618-before-kernel-change-before-seed-change-mantic-minimal-cloudimg-amd64
    * Before kernel change and before seed change
  * 
20230824-after-kernel-change-before-seed-change-mantic-minimal-cloudimg-amd64
    * After kernel change and before seed change
  * 
20230821.1-after-kernel-change-after-seed-change-mantic-minimal-cloudimg-amd64
    * After kernel change and after seed change

To manage notifications about this bug go to:
https://bugs.launchpad.net/cloud-images/+bug/2032933/+subscriptions


-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 2032933] Re: Mantic (23.10) minimal images increase in memory consumption, port usage and processes running

2023-09-04 Thread Philip Roche
@paelzer given the above findings and discussion, I would like to mark
this as Invalid for cloud-images project and continue the conversation
in the context of kernel only. +1 / -1 ?

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/2032933

Title:
  Mantic (23.10) minimal images increase in memory consumption, port
  usage and processes running

Status in cloud-images:
  Confirmed
Status in linux package in Ubuntu:
  Confirmed

Bug description:
  The Mantic (Ubuntu 23.10) download/qcow2 images available @ 
https://cloud-images.ubuntu.com/minimal/
  are undergoing some big changes prior to 23.10 release in October.

  This is a devel release so this is the perfect time to be making these
  changes but we are noticing some changes that were not expected.

  This bug is to track the unexpected changes and discuss/resolve these.

  The changes that have been made to mantic minimal:

  * Move to the linux-generic kernel from the linux-kvm kernel
    * This also involved removal of the virtio-blk driver, which is the default 
for QEMU and OpenStack, but this is being restored in an upcoming 6.5 mantic 
kernel and is being trakced @ 
https://bugs.launchpad.net/ubuntu/+source/linux/+bug/2030745
  * Move to using minimal-cloud seed - see 
https://ubuntu-archive-team.ubuntu.com/seeds/ubuntu.mantic/cloud-minimal
  * No longer installing Recommends packages
* This is during image build only and will not affect any subsequent 
package installs
  * No initramfs fallback for boot - only initramfsless boot

  The latest mantic minimal images are available @ http://cloud-
  images.ubuntu.com/minimal/daily/mantic/ and are also available in the
  public clouds.

  A package name manifest diff can be seen @
  https://pastebin.ubuntu.com/p/rRd6STnNmK/

  We have had reports of higher memory usage on an idle system, higher
  number of ports open on an idle system and higher number of process
  running on a idle system.

  To help with debugging I have built and uploaded the following images
  and package manifests to
  https://people.canonical.com/~philroche/20230824-manticl-minimal-
  LP2032933/

  * 
20230618-before-kernel-change-before-seed-change-mantic-minimal-cloudimg-amd64
    * Before kernel change and before seed change
  * 
20230824-after-kernel-change-before-seed-change-mantic-minimal-cloudimg-amd64
    * After kernel change and before seed change
  * 
20230821.1-after-kernel-change-after-seed-change-mantic-minimal-cloudimg-amd64
    * After kernel change and after seed change

To manage notifications about this bug go to:
https://bugs.launchpad.net/cloud-images/+bug/2032933/+subscriptions


-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 2032933] Re: Mantic (23.10) minimal images increase in memory consumption, port usage and processes running

2023-08-30 Thread Philip Roche
** Description changed:

- The Mantic (Ubuntu 23.10) images are undergoing some big changes prior
- to 23.10 release in October.
+ The Mantic (Ubuntu 23.10) download/qcow2 images available @ 
https://cloud-images.ubuntu.com/minimal/
+ are undergoing some big changes prior to 23.10 release in October.
  
  This is a devel release so this is the perfect time to be making these
  changes but we are noticing some changes that were not expected.
  
  This bug is to track the unexpected changes and discuss/resolve these.
  
  The changes that have been made to mantic minimal:
  
  * Move to the linux-generic kernel from the linux-kvm kernel
    * This also involved removal of the virtio-blk driver, which is the default 
for QEMU and OpenStack, but this is being restored in an upcoming 6.5 mantic 
kernel and is being trakced @ 
https://bugs.launchpad.net/ubuntu/+source/linux/+bug/2030745
  * Move to using minimal-cloud seed - see 
https://ubuntu-archive-team.ubuntu.com/seeds/ubuntu.mantic/cloud-minimal
  * No longer installing Recommends packages
+   * This is during image build only and will not affect any subsequent 
package installs
  * No initramfs fallback for boot - only initramfsless boot
  
  The latest mantic minimal images are available @ http://cloud-
  images.ubuntu.com/minimal/daily/mantic/ and are also available in the
  public clouds.
  
  A package name manifest diff can be seen @
  https://pastebin.ubuntu.com/p/rRd6STnNmK/
  
  We have had reports of higher memory usage on an idle system, higher
  number of ports open on an idle system and higher number of process
  running on a idle system.
  
  To help with debugging I have built and uploaded the following images
  and package manifests to
  https://people.canonical.com/~philroche/20230824-manticl-minimal-
  LP2032933/
  
  * 
20230618-before-kernel-change-before-seed-change-mantic-minimal-cloudimg-amd64
    * Before kernel change and before seed change
  * 
20230824-after-kernel-change-before-seed-change-mantic-minimal-cloudimg-amd64
    * After kernel change and before seed change
  * 
20230821.1-after-kernel-change-after-seed-change-mantic-minimal-cloudimg-amd64
    * After kernel change and after seed change

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/2032933

Title:
  Mantic (23.10) minimal images increase in memory consumption, port
  usage and processes running

Status in cloud-images:
  Confirmed
Status in linux package in Ubuntu:
  Confirmed

Bug description:
  The Mantic (Ubuntu 23.10) download/qcow2 images available @ 
https://cloud-images.ubuntu.com/minimal/
  are undergoing some big changes prior to 23.10 release in October.

  This is a devel release so this is the perfect time to be making these
  changes but we are noticing some changes that were not expected.

  This bug is to track the unexpected changes and discuss/resolve these.

  The changes that have been made to mantic minimal:

  * Move to the linux-generic kernel from the linux-kvm kernel
    * This also involved removal of the virtio-blk driver, which is the default 
for QEMU and OpenStack, but this is being restored in an upcoming 6.5 mantic 
kernel and is being trakced @ 
https://bugs.launchpad.net/ubuntu/+source/linux/+bug/2030745
  * Move to using minimal-cloud seed - see 
https://ubuntu-archive-team.ubuntu.com/seeds/ubuntu.mantic/cloud-minimal
  * No longer installing Recommends packages
* This is during image build only and will not affect any subsequent 
package installs
  * No initramfs fallback for boot - only initramfsless boot

  The latest mantic minimal images are available @ http://cloud-
  images.ubuntu.com/minimal/daily/mantic/ and are also available in the
  public clouds.

  A package name manifest diff can be seen @
  https://pastebin.ubuntu.com/p/rRd6STnNmK/

  We have had reports of higher memory usage on an idle system, higher
  number of ports open on an idle system and higher number of process
  running on a idle system.

  To help with debugging I have built and uploaded the following images
  and package manifests to
  https://people.canonical.com/~philroche/20230824-manticl-minimal-
  LP2032933/

  * 
20230618-before-kernel-change-before-seed-change-mantic-minimal-cloudimg-amd64
    * Before kernel change and before seed change
  * 
20230824-after-kernel-change-before-seed-change-mantic-minimal-cloudimg-amd64
    * After kernel change and before seed change
  * 
20230821.1-after-kernel-change-after-seed-change-mantic-minimal-cloudimg-amd64
    * After kernel change and after seed change

To manage notifications about this bug go to:
https://bugs.launchpad.net/cloud-images/+bug/2032933/+subscriptions


-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 2032933] Re: Mantic (23.10) minimal images increase in memory consumption, port usage and processes running

2023-08-29 Thread Philip Roche
I have uploaded further data now to
https://people.canonical.com/~philroche/20230824-manticl-minimal-
LP2032933/server-metrics/ with kernelmodules, kernelconfig, services,
timers etc. for each of the three images being inspected. This
additional data was gathered with a modified fork of the `server-test-
scripts` repo @ https://github.com/philroche/server-test-
scripts/blob/feature/local-lxc-image-execution-additional-data-
gathering/metric-server-simple/metric-server-simple.sh.

It seems that most of the mem and process increase is attributed to the
kernel change and we know that this was a conscious decision with the
following reported bugs supporting that decision.

* https://bugs.launchpad.net/ubuntu/+source/linux/+bug/2006488
* https://bugs.launchpad.net/ubuntu/+source/linux/+bug/1931841 
* https://bugs.launchpad.net/ubuntu/+source/linux/+bug/1685291

Given the above I feel what we can work on is whether any of the
process/modules introduced by the switch to the generic kernel should be
omitted for the minimal images.

The best, easiest source of this information is the data gathered from
the latest image with both the generic kernel and the switch to the new
minimal seed - https://people.canonical.com/~philroche/20230824-manticl-
minimal-LP2032933/server-metrics/20230821.1-after-kernel-change-after-
seed-change-mantic-minimal-cloudimg-amd64-data-f93870221eb8/

@seth-arnold You highlighted `ksmd`. Are there any others that concern
you.

@paelzer Are you happy to adjust your regression testing/metrics
gathering to increase the memory required knowing that it was a
conscious decision to switch kernel and incur the performance hit for
the benefit of using a kernel with more support and less reported bugs?

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/2032933

Title:
  Mantic (23.10) minimal images increase in memory consumption, port
  usage and processes running

Status in cloud-images:
  Confirmed
Status in linux package in Ubuntu:
  Confirmed

Bug description:
  The Mantic (Ubuntu 23.10) images are undergoing some big changes prior
  to 23.10 release in October.

  This is a devel release so this is the perfect time to be making these
  changes but we are noticing some changes that were not expected.

  This bug is to track the unexpected changes and discuss/resolve these.

  The changes that have been made to mantic minimal:

  * Move to the linux-generic kernel from the linux-kvm kernel
    * This also involved removal of the virtio-blk driver, which is the default 
for QEMU and OpenStack, but this is being restored in an upcoming 6.5 mantic 
kernel and is being trakced @ 
https://bugs.launchpad.net/ubuntu/+source/linux/+bug/2030745
  * Move to using minimal-cloud seed - see 
https://ubuntu-archive-team.ubuntu.com/seeds/ubuntu.mantic/cloud-minimal
  * No longer installing Recommends packages
  * No initramfs fallback for boot - only initramfsless boot

  The latest mantic minimal images are available @ http://cloud-
  images.ubuntu.com/minimal/daily/mantic/ and are also available in the
  public clouds.

  A package name manifest diff can be seen @
  https://pastebin.ubuntu.com/p/rRd6STnNmK/

  We have had reports of higher memory usage on an idle system, higher
  number of ports open on an idle system and higher number of process
  running on a idle system.

  To help with debugging I have built and uploaded the following images
  and package manifests to
  https://people.canonical.com/~philroche/20230824-manticl-minimal-
  LP2032933/

  * 
20230618-before-kernel-change-before-seed-change-mantic-minimal-cloudimg-amd64
    * Before kernel change and before seed change
  * 
20230824-after-kernel-change-before-seed-change-mantic-minimal-cloudimg-amd64
    * After kernel change and before seed change
  * 
20230821.1-after-kernel-change-after-seed-change-mantic-minimal-cloudimg-amd64
    * After kernel change and after seed change

To manage notifications about this bug go to:
https://bugs.launchpad.net/cloud-images/+bug/2032933/+subscriptions


-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 2032933] Re: Mantic (23.10) minimal images increase in memory consumption, port usage and processes running

2023-08-28 Thread Philip Roche
@paelzer

> The change of the image build sadly combined it all

See the description noting
https://people.canonical.com/~philroche/20230824-manticl-minimal-
LP2032933/ which should help in determining where the changes were
introduced as I have provided three images across the various stages of
changes - no change -> new kernel -> new kernel + new seed

For example https://pastebin.ubuntu.com/p/sJ5wGk4G7h/ show the process
diff between previous images and images with new kernel only

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/2032933

Title:
  Mantic (23.10) minimal images increase in memory consumption, port
  usage and processes running

Status in cloud-images:
  Confirmed
Status in linux package in Ubuntu:
  Confirmed

Bug description:
  The Mantic (Ubuntu 23.10) images are undergoing some big changes prior
  to 23.10 release in October.

  This is a devel release so this is the perfect time to be making these
  changes but we are noticing some changes that were not expected.

  This bug is to track the unexpected changes and discuss/resolve these.

  The changes that have been made to mantic minimal:

  * Move to the linux-generic kernel from the linux-kvm kernel
    * This also involved removal of the virtio-blk driver, which is the default 
for QEMU and OpenStack, but this is being restored in an upcoming 6.5 mantic 
kernel and is being trakced @ 
https://bugs.launchpad.net/ubuntu/+source/linux/+bug/2030745
  * Move to using minimal-cloud seed - see 
https://ubuntu-archive-team.ubuntu.com/seeds/ubuntu.mantic/cloud-minimal
  * No longer installing Recommends packages
  * No initramfs fallback for boot - only initramfsless boot

  The latest mantic minimal images are available @ http://cloud-
  images.ubuntu.com/minimal/daily/mantic/ and are also available in the
  public clouds.

  A package name manifest diff can be seen @
  https://pastebin.ubuntu.com/p/rRd6STnNmK/

  We have had reports of higher memory usage on an idle system, higher
  number of ports open on an idle system and higher number of process
  running on a idle system.

  To help with debugging I have built and uploaded the following images
  and package manifests to
  https://people.canonical.com/~philroche/20230824-manticl-minimal-
  LP2032933/

  * 
20230618-before-kernel-change-before-seed-change-mantic-minimal-cloudimg-amd64
    * Before kernel change and before seed change
  * 
20230824-after-kernel-change-before-seed-change-mantic-minimal-cloudimg-amd64
    * After kernel change and before seed change
  * 
20230821.1-after-kernel-change-after-seed-change-mantic-minimal-cloudimg-amd64
    * After kernel change and after seed change

To manage notifications about this bug go to:
https://bugs.launchpad.net/cloud-images/+bug/2032933/+subscriptions


-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 2032933] Re: Mantic (23.10) minimal images increase in memory consumption, port usage and processes running

2023-08-28 Thread Philip Roche
The diff in process count from kernel change image -> kernel change +
seed change image is actually a reduction in processes - see diff @
https://pastebin.ubuntu.com/p/PXtQM9gB2K/

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/2032933

Title:
  Mantic (23.10) minimal images increase in memory consumption, port
  usage and processes running

Status in cloud-images:
  Confirmed
Status in linux package in Ubuntu:
  Confirmed

Bug description:
  The Mantic (Ubuntu 23.10) images are undergoing some big changes prior
  to 23.10 release in October.

  This is a devel release so this is the perfect time to be making these
  changes but we are noticing some changes that were not expected.

  This bug is to track the unexpected changes and discuss/resolve these.

  The changes that have been made to mantic minimal:

  * Move to the linux-generic kernel from the linux-kvm kernel
    * This also involved removal of the virtio-blk driver, which is the default 
for QEMU and OpenStack, but this is being restored in an upcoming 6.5 mantic 
kernel and is being trakced @ 
https://bugs.launchpad.net/ubuntu/+source/linux/+bug/2030745
  * Move to using minimal-cloud seed - see 
https://ubuntu-archive-team.ubuntu.com/seeds/ubuntu.mantic/cloud-minimal
  * No longer installing Recommends packages
  * No initramfs fallback for boot - only initramfsless boot

  The latest mantic minimal images are available @ http://cloud-
  images.ubuntu.com/minimal/daily/mantic/ and are also available in the
  public clouds.

  A package name manifest diff can be seen @
  https://pastebin.ubuntu.com/p/rRd6STnNmK/

  We have had reports of higher memory usage on an idle system, higher
  number of ports open on an idle system and higher number of process
  running on a idle system.

  To help with debugging I have built and uploaded the following images
  and package manifests to
  https://people.canonical.com/~philroche/20230824-manticl-minimal-
  LP2032933/

  * 
20230618-before-kernel-change-before-seed-change-mantic-minimal-cloudimg-amd64
    * Before kernel change and before seed change
  * 
20230824-after-kernel-change-before-seed-change-mantic-minimal-cloudimg-amd64
    * After kernel change and before seed change
  * 
20230821.1-after-kernel-change-after-seed-change-mantic-minimal-cloudimg-amd64
    * After kernel change and after seed change

To manage notifications about this bug go to:
https://bugs.launchpad.net/cloud-images/+bug/2032933/+subscriptions


-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1882955] Re: LXD 4.2 broken on linux-kvm due to missing VLAN filtering

2020-06-23 Thread Philip Roche
CPC are seeing this issue in _all_ minimal cloud images testing with LXD
snap version 4.2 or greater. This blocks promotion of all minimal cloud
download images and blocks build and publication of both daily and
release cloud images.

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux-kvm in Ubuntu.
https://bugs.launchpad.net/bugs/1882955

Title:
  LXD 4.2 broken on linux-kvm due to missing VLAN filtering

Status in linux-kvm package in Ubuntu:
  Triaged

Bug description:
  This is another case of linux-kvm having unexplained differences
  compared to linux-generic in areas that aren't related to hardware
  drivers (see other bug we filed for missing nft).

  This time, CPC is reporting that LXD no longer works on linux-kvm as
  we now set vlan filtering on our bridges to prevent containers from
  escaping firewalling through custom vlan tags.

  This relies on CONFIG_BRIDGE_VLAN_FILTERING which is a built-in on the
  generic kernel but is apparently missing on linux-kvm (I don't have
  any system running that kernel to confirm its config, but the behavior
  certainly matches that).

  We need this fixed in focal and groovy.

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/linux-kvm/+bug/1882955/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1861704] Re: ubuntu-fan recommends netcat package in universe

2020-02-03 Thread Philip Roche
It appears that a bug was already filed against netcat

https://bugs.launchpad.net/ubuntu/+source/netcat-openbsd/+bug/1780316

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to ubuntu-fan in Ubuntu.
https://bugs.launchpad.net/bugs/1861704

Title:
  ubuntu-fan recommends netcat package  in universe

Status in ubuntu-fan package in Ubuntu:
  New

Bug description:
  Following some debug of the docker.io package in universe, we
  (Canonical CPC) discovered that the ubuntu-fan package from main and
  the netcat-traditional package from universe were being installed.

  This was due to the following dependency/recommends tree:

  * docker.io (in universe) recommends ubuntu-fan (main)
  * ubuntu-fan (main) recommends netcat (universe)
  * netcat (universe) is a transitional package and depends on 
netcat-traditional (universe)

  Our concern is that this might be a packaging violation as ubuntu-fan
  is recommending a package not in main.

  > In addition, the packages in main
  > 
  > must not require a package outside of main for compilation or execution 
(thus, the package must 
  > not declare a "Depends", "Recommends", or "Build-Depends" relationship on a 
non-main package),

  Source: https://people.canonical.com/~cjwatson/ubuntu-
  policy/policy.html/ch-archive.html#s-main

  I will file a bug against netcat too to start a discussion on netcat
  being built from netcat-openbsd (main) instead of netcat-traditional
  (universe).

  
  Our feeling is that netcat is such a frequently depended on or recommended 
package that it being present in main would benefit Ubuntu.

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/ubuntu-fan/+bug/1861704/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1861704] [NEW] ubuntu-fan recommends netcat package in universe

2020-02-03 Thread Philip Roche
Public bug reported:

Following some debug of the docker.io package in universe, we (Canonical
CPC) discovered that the ubuntu-fan package from main and the netcat-
traditional package from universe were being installed.

This was due to the following dependency/recommends tree:

* docker.io (in universe) recommends ubuntu-fan (main)
* ubuntu-fan (main) recommends netcat (universe)
* netcat (universe) is a transitional package and depends on netcat-traditional 
(universe)

Our concern is that this might be a packaging violation as ubuntu-fan is
recommending a package not in main.

> In addition, the packages in main
> 
> must not require a package outside of main for compilation or execution 
> (thus, the package must 
> not declare a "Depends", "Recommends", or "Build-Depends" relationship on a 
> non-main package),

Source: https://people.canonical.com/~cjwatson/ubuntu-policy/policy.html
/ch-archive.html#s-main

I will file a bug against netcat too to start a discussion on netcat
being built from netcat-openbsd (main) instead of netcat-traditional
(universe).


Our feeling is that netcat is such a frequently depended on or recommended 
package that it being present in main would benefit Ubuntu.

** Affects: ubuntu-fan (Ubuntu)
 Importance: Undecided
 Status: New

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to ubuntu-fan in Ubuntu.
https://bugs.launchpad.net/bugs/1861704

Title:
  ubuntu-fan recommends netcat package  in universe

Status in ubuntu-fan package in Ubuntu:
  New

Bug description:
  Following some debug of the docker.io package in universe, we
  (Canonical CPC) discovered that the ubuntu-fan package from main and
  the netcat-traditional package from universe were being installed.

  This was due to the following dependency/recommends tree:

  * docker.io (in universe) recommends ubuntu-fan (main)
  * ubuntu-fan (main) recommends netcat (universe)
  * netcat (universe) is a transitional package and depends on 
netcat-traditional (universe)

  Our concern is that this might be a packaging violation as ubuntu-fan
  is recommending a package not in main.

  > In addition, the packages in main
  > 
  > must not require a package outside of main for compilation or execution 
(thus, the package must 
  > not declare a "Depends", "Recommends", or "Build-Depends" relationship on a 
non-main package),

  Source: https://people.canonical.com/~cjwatson/ubuntu-
  policy/policy.html/ch-archive.html#s-main

  I will file a bug against netcat too to start a discussion on netcat
  being built from netcat-openbsd (main) instead of netcat-traditional
  (universe).

  
  Our feeling is that netcat is such a frequently depended on or recommended 
package that it being present in main would benefit Ubuntu.

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/ubuntu-fan/+bug/1861704/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1795017] Re: linux-azure on cosmic minimal image boot speed

2018-10-11 Thread Philip Roche
Myself and @kamalmostafa spoke and the issue is not obvious especially
as linux-azure is installed in the base image too which does not exhibit
the issue.

Although not obvious @kamalmostafa did suspect the issue to lie
somewhere with random number generation and a related package/kernel
patch.

One line of investigation is to check the package diff between the
minimal image with initramfs-tools installed and the image without that
installed. That diff is @ https://pastebin.canonical.com/p/7hJTSQHbyD/

Another possibility is to check the package diff between base and
minimal. That diff is @ https://pastebin.canonical.com/p/B9m7GbsFhs/

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux-azure in Ubuntu.
https://bugs.launchpad.net/bugs/1795017

Title:
  linux-azure on cosmic minimal image boot speed

Status in linux-azure package in Ubuntu:
  New

Bug description:
  Hi,

  We're planning to share a new Cosmic minimal image for Azure including
  the linux-azure custom kernel.

  I have found boot speed increase from ~55s in non minimal image to
  over 6mins for minimal image.

  I have gathered boot log, systemd logs and cloud-init logs for this
  instance @ https://private-fileshare.canonical.com/~philroche/azure-
  cosmic-20180927/minimal/. You can see the same logs for the non
  minimal image @ https://private-fileshare.canonical.com/~philroche
  /azure-cosmic-20180927/base

  Most of the time seems to be spent in crng and "task swapper/0:1" (?)

  [5.774760] sd 3:0:0:0: [sdc] Attached SCSI disk
  [   71.388989] random: crng init done
  [  242.656057] INFO: task swapper/0:1 blocked for more than 120 seconds.
  [  242.688434]   Not tainted 4.15.0-1018-azure #18-Ubuntu

  Source: https://private-fileshare.canonical.com/~philroche/azure-
  cosmic-20180927/minimal/iengal3seeSee-vm-imagedebug-boot.log

  Please let me know if you need any more information or would like me
  to launch an instance for testing.

  
  I understand that minimal images are intended to boot without initramfs but I 
did install initramfs-tools in another test image which did result in much 
better boot times.

  See https://private-fileshare.canonical.com/~philroche/azure-
  cosmic-20180928/ for logs.

  Minimal image boot time 48.784s

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/linux-azure/+bug/1795017/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1795017] Re: linux-azure on cosmic minimal image boot speed

2018-10-11 Thread Philip Roche
@kamalmostafa @mhcerri

I noticed that linux-azure 4.18.0.1003.3 was uploaded to cosmic-proposed
so I built a new test image.

For convenience I have imported your SSH keys to 'ssh
ubuntu@51.140.217.58'.

All logs and VHD etc. have been uploaded to https://private-
fileshare.canonical.com/~philroche/azure-cosmic-20181011/

** Attachment added: "cosmic minimal image with linux-azure 4.18.0.1003.3 (from 
cosmic-proposed) boot log"
   
https://bugs.launchpad.net/ubuntu/+source/linux-azure/+bug/1795017/+attachment/5199957/+files/eeDaquoh3aiqu-vm-imagedebug-boot.log

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux-azure in Ubuntu.
https://bugs.launchpad.net/bugs/1795017

Title:
  linux-azure on cosmic minimal image boot speed

Status in linux-azure package in Ubuntu:
  New

Bug description:
  Hi,

  We're planning to share a new Cosmic minimal image for Azure including
  the linux-azure custom kernel.

  I have found boot speed increase from ~55s in non minimal image to
  over 6mins for minimal image.

  I have gathered boot log, systemd logs and cloud-init logs for this
  instance @ https://private-fileshare.canonical.com/~philroche/azure-
  cosmic-20180927/minimal/. You can see the same logs for the non
  minimal image @ https://private-fileshare.canonical.com/~philroche
  /azure-cosmic-20180927/base

  Most of the time seems to be spent in crng and "task swapper/0:1" (?)

  [5.774760] sd 3:0:0:0: [sdc] Attached SCSI disk
  [   71.388989] random: crng init done
  [  242.656057] INFO: task swapper/0:1 blocked for more than 120 seconds.
  [  242.688434]   Not tainted 4.15.0-1018-azure #18-Ubuntu

  Source: https://private-fileshare.canonical.com/~philroche/azure-
  cosmic-20180927/minimal/iengal3seeSee-vm-imagedebug-boot.log

  Please let me know if you need any more information or would like me
  to launch an instance for testing.

  
  I understand that minimal images are intended to boot without initramfs but I 
did install initramfs-tools in another test image which did result in much 
better boot times.

  See https://private-fileshare.canonical.com/~philroche/azure-
  cosmic-20180928/ for logs.

  Minimal image boot time 48.784s

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/linux-azure/+bug/1795017/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1795017] Re: linux-azure on cosmic minimal image boot speed

2018-10-01 Thread Philip Roche
On request on IRC I have tried a new image with linux-azure
4.18.0.1002.2 from cosmic-proposed.

This too exhibited the slow boot

5min 19.070s (kernel) + 41.724s (userspace) = 6min 794ms

Full logs @ https://private-fileshare.canonical.com/~philroche/azure-
cosmic-20181001/

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux-azure in Ubuntu.
https://bugs.launchpad.net/bugs/1795017

Title:
  linux-azure on cosmic minimal image boot speed

Status in linux-azure package in Ubuntu:
  New

Bug description:
  Hi,

  We're planning to share a new Cosmic minimal image for Azure including
  the linux-azure custom kernel.

  I have found boot speed increase from ~55s in non minimal image to
  over 6mins for minimal image.

  I have gathered boot log, systemd logs and cloud-init logs for this
  instance @ https://private-fileshare.canonical.com/~philroche/azure-
  cosmic-20180927/minimal/. You can see the same logs for the non
  minimal image @ https://private-fileshare.canonical.com/~philroche
  /azure-cosmic-20180927/base

  Most of the time seems to be spent in crng and "task swapper/0:1" (?)

  [5.774760] sd 3:0:0:0: [sdc] Attached SCSI disk
  [   71.388989] random: crng init done
  [  242.656057] INFO: task swapper/0:1 blocked for more than 120 seconds.
  [  242.688434]   Not tainted 4.15.0-1018-azure #18-Ubuntu

  Source: https://private-fileshare.canonical.com/~philroche/azure-
  cosmic-20180927/minimal/iengal3seeSee-vm-imagedebug-boot.log

  Please let me know if you need any more information or would like me
  to launch an instance for testing.

  
  I understand that minimal images are intended to boot without initramfs but I 
did install initramfs-tools in another test image which did result in much 
better boot times.

  See https://private-fileshare.canonical.com/~philroche/azure-
  cosmic-20180928/ for logs.

  Minimal image boot time 48.784s

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/linux-azure/+bug/1795017/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1795017] Re: linux-azure on cosmic minimal image boot speed

2018-10-01 Thread Philip Roche
** Attachment added: "cosmic minimal image with linux-azure 4.18.0.1002.2 (from 
cosmic-proposed) boot log"
   
https://bugs.launchpad.net/ubuntu/+source/linux-azure/+bug/1795017/+attachment/5195337/+files/chohv4Va7Ahx7-vm-imagedebug-boot.log

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux-azure in Ubuntu.
https://bugs.launchpad.net/bugs/1795017

Title:
  linux-azure on cosmic minimal image boot speed

Status in linux-azure package in Ubuntu:
  New

Bug description:
  Hi,

  We're planning to share a new Cosmic minimal image for Azure including
  the linux-azure custom kernel.

  I have found boot speed increase from ~55s in non minimal image to
  over 6mins for minimal image.

  I have gathered boot log, systemd logs and cloud-init logs for this
  instance @ https://private-fileshare.canonical.com/~philroche/azure-
  cosmic-20180927/minimal/. You can see the same logs for the non
  minimal image @ https://private-fileshare.canonical.com/~philroche
  /azure-cosmic-20180927/base

  Most of the time seems to be spent in crng and "task swapper/0:1" (?)

  [5.774760] sd 3:0:0:0: [sdc] Attached SCSI disk
  [   71.388989] random: crng init done
  [  242.656057] INFO: task swapper/0:1 blocked for more than 120 seconds.
  [  242.688434]   Not tainted 4.15.0-1018-azure #18-Ubuntu

  Source: https://private-fileshare.canonical.com/~philroche/azure-
  cosmic-20180927/minimal/iengal3seeSee-vm-imagedebug-boot.log

  Please let me know if you need any more information or would like me
  to launch an instance for testing.

  
  I understand that minimal images are intended to boot without initramfs but I 
did install initramfs-tools in another test image which did result in much 
better boot times.

  See https://private-fileshare.canonical.com/~philroche/azure-
  cosmic-20180928/ for logs.

  Minimal image boot time 48.784s

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/linux-azure/+bug/1795017/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1795017] [NEW] linux-azure on cosmic minimal image boot speed

2018-09-28 Thread Philip Roche
Public bug reported:

Hi,

We're planning to share a new Cosmic minimal image for Azure including
the linux-azure custom kernel.

I have found boot speed increase from ~55s in non minimal image to over
6mins for minimal image.

I have gathered boot log, systemd logs and cloud-init logs for this
instance @ https://private-fileshare.canonical.com/~philroche/azure-
cosmic-20180927/minimal/. You can see the same logs for the non minimal
image @ https://private-fileshare.canonical.com/~philroche/azure-
cosmic-20180927/base

Most of the time seems to be spent in crng and "task swapper/0:1" (?)

[5.774760] sd 3:0:0:0: [sdc] Attached SCSI disk
[   71.388989] random: crng init done
[  242.656057] INFO: task swapper/0:1 blocked for more than 120 seconds.
[  242.688434]   Not tainted 4.15.0-1018-azure #18-Ubuntu

Source: https://private-fileshare.canonical.com/~philroche/azure-
cosmic-20180927/minimal/iengal3seeSee-vm-imagedebug-boot.log

Please let me know if you need any more information or would like me to
launch an instance for testing.


I understand that minimal images are intended to boot without initramfs but I 
did install initramfs-tools in another test image which did result in much 
better boot times.

See https://private-fileshare.canonical.com/~philroche/azure-
cosmic-20180928/ for logs.

Minimal image boot time 48.784s

** Affects: linux-azure (Ubuntu)
 Importance: Undecided
 Status: New

** Attachment added: "cosmic minimal image with linux-azure boot log"
   
https://bugs.launchpad.net/bugs/1795017/+attachment/5194121/+files/iengal3seeSee-vm-imagedebug-boot.log

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux-azure in Ubuntu.
https://bugs.launchpad.net/bugs/1795017

Title:
  linux-azure on cosmic minimal image boot speed

Status in linux-azure package in Ubuntu:
  New

Bug description:
  Hi,

  We're planning to share a new Cosmic minimal image for Azure including
  the linux-azure custom kernel.

  I have found boot speed increase from ~55s in non minimal image to
  over 6mins for minimal image.

  I have gathered boot log, systemd logs and cloud-init logs for this
  instance @ https://private-fileshare.canonical.com/~philroche/azure-
  cosmic-20180927/minimal/. You can see the same logs for the non
  minimal image @ https://private-fileshare.canonical.com/~philroche
  /azure-cosmic-20180927/base

  Most of the time seems to be spent in crng and "task swapper/0:1" (?)

  [5.774760] sd 3:0:0:0: [sdc] Attached SCSI disk
  [   71.388989] random: crng init done
  [  242.656057] INFO: task swapper/0:1 blocked for more than 120 seconds.
  [  242.688434]   Not tainted 4.15.0-1018-azure #18-Ubuntu

  Source: https://private-fileshare.canonical.com/~philroche/azure-
  cosmic-20180927/minimal/iengal3seeSee-vm-imagedebug-boot.log

  Please let me know if you need any more information or would like me
  to launch an instance for testing.

  
  I understand that minimal images are intended to boot without initramfs but I 
did install initramfs-tools in another test image which did result in much 
better boot times.

  See https://private-fileshare.canonical.com/~philroche/azure-
  cosmic-20180928/ for logs.

  Minimal image boot time 48.784s

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/linux-azure/+bug/1795017/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1779961] Re: Boot delays with 4.15.0-24 based kernels in images

2018-07-04 Thread Philip Roche
The meta-packages have been rolled back and the fix is in progress of
being applied and uploaded.

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1779961

Title:
  Boot delays with 4.15.0-24 based kernels in images

Status in linux package in Ubuntu:
  Triaged

Bug description:
  We (Canonical CPC) are seeing an issue with today's build of the
  bionic minimal cloud images. Boot stops at "Starting Initial cloud-
  init job (metadata service crawler)..." for nearly 5 minutes (booting
  on scalingstack and locally with kvm) this is with a new kernel, but
  otherwise no changes over the prior image.

  Image (manifest and logs) exhibiting this behaviour @ https://private-
  fileshare.canonical.com/~philroche/bionic-minimal-20180703/

  With server teams's help we have narrowed it down to "Synchronized to
  time server 91.189.91.157:123 (ntp.ubuntu.com)." taking over 4
  minutes.

  On Kamal's instruction we also tried  linux-kvm 4.15.0-1013.13 and
  bionic GA linux (4.15.0-24.26). Both of which also exhibit this
  regression.

  For reference: Kamal noted on IRC that the problem was possibly
  introduced in the GA kernel and is not specific to the kvm kernel.

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/linux/+bug/1779961/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp


[Kernel-packages] [Bug 1779961] [NEW] Bionic minimal image boot hang with linux-image-4.15.0-1012-kvm kernel

2018-07-03 Thread Philip Roche
Public bug reported:

We (Canonical CPC) are seeing an issue with today's build of the bionic
minimal cloud images. Boot stops at "Starting Initial cloud-init job
(metadata service crawler)..." for nearly 5 minutes (booting on
scalingstack and locally with kvm) this is with a new kernel, but
otherwise no changes over the prior image.

Image (manifest and logs) exhibiting this behaviour @ https://private-
fileshare.canonical.com/~philroche/bionic-minimal-20180703/

With server teams's help we have narrowed it down to "Synchronized to
time server 91.189.91.157:123 (ntp.ubuntu.com)." taking over 4 minutes.

On Kamal's instruction we also tried  linux-kvm 4.15.0-1013.13 and
bionic GA linux (4.15.0-24.26). Both of which also exhibit this
regression.

For reference: Kamal noted on IRC that the problem was possibly
introduced in the GA kernel and is not specific to the kvm kernel.

** Affects: linux (Ubuntu)
 Importance: Undecided
 Status: Incomplete


** Tags: bionic

-- 
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1779961

Title:
  Bionic minimal image boot hang with  linux-image-4.15.0-1012-kvm
  kernel

Status in linux package in Ubuntu:
  Incomplete

Bug description:
  We (Canonical CPC) are seeing an issue with today's build of the
  bionic minimal cloud images. Boot stops at "Starting Initial cloud-
  init job (metadata service crawler)..." for nearly 5 minutes (booting
  on scalingstack and locally with kvm) this is with a new kernel, but
  otherwise no changes over the prior image.

  Image (manifest and logs) exhibiting this behaviour @ https://private-
  fileshare.canonical.com/~philroche/bionic-minimal-20180703/

  With server teams's help we have narrowed it down to "Synchronized to
  time server 91.189.91.157:123 (ntp.ubuntu.com)." taking over 4
  minutes.

  On Kamal's instruction we also tried  linux-kvm 4.15.0-1013.13 and
  bionic GA linux (4.15.0-24.26). Both of which also exhibit this
  regression.

  For reference: Kamal noted on IRC that the problem was possibly
  introduced in the GA kernel and is not specific to the kvm kernel.

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/linux/+bug/1779961/+subscriptions

-- 
Mailing list: https://launchpad.net/~kernel-packages
Post to : kernel-packages@lists.launchpad.net
Unsubscribe : https://launchpad.net/~kernel-packages
More help   : https://help.launchpad.net/ListHelp