[Bug 1850540] Re: multi-zone raid0 corruption

2021-11-17 Thread Brian Murray
** Changed in: ubuntu-release-notes
   Status: New => Won't Fix

-- 
You received this bug notification because you are a member of Ubuntu
Bugs, which is subscribed to Ubuntu.
https://bugs.launchpad.net/bugs/1850540

Title:
  multi-zone raid0 corruption

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu-release-notes/+bug/1850540/+subscriptions


-- 
ubuntu-bugs mailing list
ubuntu-bugs@lists.ubuntu.com
https://lists.ubuntu.com/mailman/listinfo/ubuntu-bugs

[Bug 1850540] Re: multi-zone raid0 corruption

2021-10-13 Thread Steve Langasek
The Precise Pangolin has reached end of life, so this bug will not be
fixed for that release

** Changed in: mdadm (Ubuntu Precise)
   Status: New => Won't Fix

-- 
You received this bug notification because you are a member of Ubuntu
Bugs, which is subscribed to Ubuntu.
https://bugs.launchpad.net/bugs/1850540

Title:
  multi-zone raid0 corruption

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu-release-notes/+bug/1850540/+subscriptions


-- 
ubuntu-bugs mailing list
ubuntu-bugs@lists.ubuntu.com
https://lists.ubuntu.com/mailman/listinfo/ubuntu-bugs

[Bug 1850540] Re: multi-zone raid0 corruption

2021-10-13 Thread Steve Langasek
The Precise Pangolin has reached end of life, so this bug will not be
fixed for that release

** Changed in: linux (Ubuntu Precise)
   Status: New => Won't Fix

-- 
You received this bug notification because you are a member of Ubuntu
Bugs, which is subscribed to Ubuntu.
https://bugs.launchpad.net/bugs/1850540

Title:
  multi-zone raid0 corruption

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu-release-notes/+bug/1850540/+subscriptions


-- 
ubuntu-bugs mailing list
ubuntu-bugs@lists.ubuntu.com
https://lists.ubuntu.com/mailman/listinfo/ubuntu-bugs

[Bug 1850540] Re: multi-zone raid0 corruption

2020-07-02 Thread Steve Langasek
** Changed in: linux (Ubuntu Disco)
   Status: Fix Committed => Won't Fix

** Changed in: mdadm (Ubuntu Disco)
   Status: Fix Committed => Won't Fix

-- 
You received this bug notification because you are a member of Ubuntu
Bugs, which is subscribed to Ubuntu.
https://bugs.launchpad.net/bugs/1850540

Title:
  multi-zone raid0 corruption

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu-release-notes/+bug/1850540/+subscriptions

-- 
ubuntu-bugs mailing list
ubuntu-bugs@lists.ubuntu.com
https://lists.ubuntu.com/mailman/listinfo/ubuntu-bugs

[Bug 1850540] Re: multi-zone raid0 corruption

2020-03-16 Thread Launchpad Bug Tracker
This bug was fixed in the package linux - 5.4.0-18.22

---
linux (5.4.0-18.22) focal; urgency=medium

  * focal/linux: 5.4.0-18.22 -proposed tracker (LP: #1866488)

  * Packaging resync (LP: #1786013)
- [Packaging] resync getabis
- [Packaging] update helper scripts

  * Add sysfs attribute to show remapped NVMe (LP: #1863621)
- SAUCE: ata: ahci: Add sysfs attribute to show remapped NVMe device count

  * [20.04 FEAT] Compression improvements in Linux kernel (LP: #1830208)
- lib/zlib: add s390 hardware support for kernel zlib_deflate
- s390/boot: rename HEAP_SIZE due to name collision
- lib/zlib: add s390 hardware support for kernel zlib_inflate
- s390/boot: add dfltcc= kernel command line parameter
- lib/zlib: add zlib_deflate_dfltcc_enabled() function
- btrfs: use larger zlib buffer for s390 hardware compression
- [Config] Introducing s390x specific kernel config option 
CONFIG_ZLIB_DFLTCC

  * [UBUNTU 20.04] s390x/pci: increase CONFIG_PCI_NR_FUNCTIONS to 512 in kernel
config (LP: #1866056)
- [Config] Increase CONFIG_PCI_NR_FUNCTIONS from 64 to 512 starting with 
focal
  on s390x

  * CONFIG_IP_MROUTE_MULTIPLE_TABLES is not set (LP: #1865332)
- [Config] CONFIG_IP_MROUTE_MULTIPLE_TABLES=y

  * Dell XPS 13 9300 Intel 1650S wifi [34f0:1651] fails to load firmware
(LP: #1865962)
- iwlwifi: remove IWL_DEVICE_22560/IWL_DEVICE_FAMILY_22560
- iwlwifi: 22000: fix some indentation
- iwlwifi: pcie: rx: use rxq queue_size instead of constant
- iwlwifi: allocate more receive buffers for HE devices
- iwlwifi: remove some outdated iwl22000 configurations
- iwlwifi: assume the driver_data is a trans_cfg, but allow full cfg

  * [FOCAL][REGRESSION] Intel Gen 9 brightness cannot be controlled
(LP: #1861521)
- Revert "USUNTU: SAUCE: drm/i915: Force DPCD backlight mode on Dell 
Precision
  4K sku"
- Revert "UBUNTU: SAUCE: drm/i915: Force DPCD backlight mode on X1 Extreme 
2nd
  Gen 4K AMOLED panel"
- SAUCE: drm/dp: Introduce EDID-based quirks
- SAUCE: drm/i915: Force DPCD backlight mode on X1 Extreme 2nd Gen 4K AMOLED
  panel
- SAUCE: drm/i915: Force DPCD backlight mode for some Dell CML 2020 panels

  * [20.04 FEAT] Enable proper kprobes on ftrace support (LP: #1865858)
- s390/ftrace: save traced function caller
- s390: support KPROBES_ON_FTRACE

  * alsa/sof: load different firmware on different platforms (LP: #1857409)
- ASoC: SOF: Intel: hda: use fallback for firmware name
- ASoC: Intel: acpi-match: split CNL tables in three
- ASoC: SOF: Intel: Fix CFL and CML FW nocodec binary names.

  * [UBUNTU 20.04] Enable CONFIG_NET_SWITCHDEV in kernel config for s390x
starting with focal (LP: #1865452)
- [Config] Enable CONFIG_NET_SWITCHDEV in kernel config for s390x starting
  with focal

  * Focal update: v5.4.24 upstream stable release (LP: #1866333)
- io_uring: grab ->fs as part of async offload
- EDAC: skx_common: downgrade message importance on missing PCI device
- net: dsa: b53: Ensure the default VID is untagged
- net: fib_rules: Correctly set table field when table number exceeds 8 bits
- net: macb: ensure interface is not suspended on at91rm9200
- net: mscc: fix in frame extraction
- net: phy: restore mdio regs in the iproc mdio driver
- net: sched: correct flower port blocking
- net/tls: Fix to avoid gettig invalid tls record
- nfc: pn544: Fix occasional HW initialization failure
- qede: Fix race between rdma destroy workqueue and link change event
- Revert "net: dev: introduce support for sch BYPASS for lockless qdisc"
- udp: rehash on disconnect
- sctp: move the format error check out of __sctp_sf_do_9_1_abort
- bnxt_en: Improve device shutdown method.
- bnxt_en: Issue PCIe FLR in kdump kernel to cleanup pending DMAs.
- bonding: add missing netdev_update_lockdep_key()
- net: export netdev_next_lower_dev_rcu()
- bonding: fix lockdep warning in bond_get_stats()
- ipv6: Fix route replacement with dev-only route
- ipv6: Fix nlmsg_flags when splitting a multipath route
- ipmi:ssif: Handle a possible NULL pointer reference
- drm/msm: Set dma maximum segment size for mdss
- sched/core: Don't skip remote tick for idle CPUs
- timers/nohz: Update NOHZ load in remote tick
- sched/fair: Prevent unlimited runtime on throttled group
- dax: pass NOWAIT flag to iomap_apply
- mac80211: consider more elements in parsing CRC
- cfg80211: check wiphy driver existence for drvinfo report
- s390/zcrypt: fix card and queue total counter wrap
- qmi_wwan: re-add DW5821e pre-production variant
- qmi_wwan: unconditionally reject 2 ep interfaces
- NFSv4: Fix races between open and dentry revalidation
- perf/smmuv3: Use platform_get_irq_optional() for wired interrupt
- perf/x86/intel: Add Elkhart Lake support
- perf/x86/cstate: Add Tremont 

[Bug 1850540] Re: multi-zone raid0 corruption

2020-02-17 Thread Launchpad Bug Tracker
This bug was fixed in the package linux - 4.4.0-174.204

---
linux (4.4.0-174.204) xenial; urgency=medium

  * xenial/linux: 4.4.0-174.204 -proposed tracker (LP: #1861122)

  * Xenial update: 4.4.211 upstream stable release (LP: #1860681)
- hidraw: Return EPOLLOUT from hidraw_poll
- HID: hidraw: Fix returning EPOLLOUT from hidraw_poll
- HID: hidraw, uhid: Always report EPOLLOUT
- cfg80211/mac80211: make ieee80211_send_layer2_update a public function
- mac80211: Do not send Layer 2 Update frame before authorization
- media: usb:zr364xx:Fix KASAN:null-ptr-deref Read in 
zr364xx_vidioc_querycap
- p54usb: Fix race between disconnect and firmware loading
- ALSA: line6: Fix write on zero-sized buffer
- ALSA: line6: Fix memory leak at line6_init_pcm() error path
- xen: let alloc_xenballooned_pages() fail if not enough memory free
- wimax: i2400: fix memory leak
- wimax: i2400: Fix memory leak in i2400m_op_rfkill_sw_toggle
- ext4: fix use-after-free race with debug_want_extra_isize
- ext4: add more paranoia checking in ext4_expand_extra_isize handling
- rtc: mt6397: fix alarm register overwrite
- iommu: Remove device link to group on failure
- gpio: Fix error message on out-of-range GPIO in lookup table
- hsr: reset network header when supervision frame is created
- cifs: Adjust indentation in smb2_open_file
- RDMA/srpt: Report the SCSI residual to the initiator
- scsi: enclosure: Fix stale device oops with hot replug
- scsi: sd: Clear sdkp->protection_type if disk is reformatted without PI
- platform/x86: asus-wmi: Fix keyboard brightness cannot be set to 0
- iio: imu: adis16480: assign bias value only if operation succeeded
- mei: fix modalias documentation
- clk: samsung: exynos5420: Preserve CPU clocks configuration during
  suspend/resume
- compat_ioctl: handle SIOCOUTQNSD
- tty: serial: imx: use the sg count from dma_map_sg
- tty: serial: pch_uart: correct usage of dma_unmap_sg
- media: exynos4-is: Fix recursive locking in isp_video_release()
- spi: atmel: fix handling of cs_change set on non-last xfer
- rtlwifi: Remove unnecessary NULL check in rtl_regd_init
- rtc: msm6242: Fix reading of 10-hour digit
- rseq/selftests: Turn off timeout setting
- hexagon: work around compiler crash
- ocfs2: call journal flush to mark journal as empty after journal recovery
  when mount
- ALSA: seq: Fix racy access for queue timer in proc read
- Fix built-in early-load Intel microcode alignment
- block: fix an integer overflow in logical block size
- USB: serial: simple: Add Motorola Solutions TETRA MTP3xxx and MTP85xx
- USB: serial: opticon: fix control-message timeouts
- USB: serial: suppress driver bind attributes
- USB: serial: ch341: handle unbound port at reset_resume
- USB: serial: io_edgeport: add missing active-port sanity check
- USB: serial: quatech2: handle unbound ports
- scsi: mptfusion: Fix double fetch bug in ioctl
- usb: core: hub: Improved device recognition on remote wakeup
- x86/efistub: Disable paging at mixed mode entry
- mm/page-writeback.c: avoid potential division by zero in 
wb_min_max_ratio()
- net: stmmac: 16KB buffer must be 16 byte aligned
- net: stmmac: Enable 16KB buffer size
- USB: serial: io_edgeport: use irqsave() in USB's complete callback
- USB: serial: io_edgeport: handle unbound ports on URB completion
- USB: serial: keyspan: handle unbound ports
- scsi: fnic: use kernel's '%pM' format option to print MAC
- scsi: fnic: fix invalid stack access
- arm64: dts: agilex/stratix10: fix pmu interrupt numbers
- netfilter: fix a use-after-free in mtype_destroy()
- batman-adv: Fix DAT candidate selection on little endian systems
- macvlan: use skb_reset_mac_header() in macvlan_queue_xmit()
- r8152: add missing endpoint sanity check
- tcp: fix marked lost packets not being retransmitted
- net: usb: lan78xx: limit size of local TSO packets
- xen/blkfront: Adjust indentation in xlvbd_alloc_gendisk
- cw1200: Fix a signedness bug in cw1200_load_firmware()
- cfg80211: check for set_wiphy_params
- scsi: esas2r: unlock on error in esas2r_nvram_read_direct()
- scsi: qla4xxx: fix double free bug
- scsi: bnx2i: fix potential use after free
- scsi: target: core: Fix a pr_debug() argument
- scsi: core: scsi_trace: Use get_unaligned_be*()
- perf probe: Fix wrong address verification
- regulator: ab8500: Remove SYSCLKREQ from enum ab8505_regulator_id
- Linux 4.4.211

  * Xenial update: 4.4.210 upstream stable release (LP: #1859865)
- chardev: Avoid potential use-after-free in 'chrdev_open()'
- usb: chipidea: host: Disable port power only if previously enabled
- ALSA: usb-audio: Apply the sample rate quirk for Bose Companion 5
- kernel/trace: Fix do not unregister tracepoints when 

[Bug 1850540] Re: multi-zone raid0 corruption

2020-02-17 Thread Launchpad Bug Tracker
This bug was fixed in the package linux - 4.15.0-88.88

---
linux (4.15.0-88.88) bionic; urgency=medium

  * bionic/linux: 4.15.0-88.88 -proposed tracker (LP: #1862824)

  * Segmentation fault (kernel oops) with memory-hotplug in
ubuntu_kernel_selftests on Bionic kernel (LP: #1862312)
- Revert "mm/memory_hotplug: fix online/offline_pages called w.o.
  mem_hotplug_lock"
- mm/memory_hotplug: fix online/offline_pages called w.o. mem_hotplug_lock

linux (4.15.0-87.87) bionic; urgency=medium

  * bionic/linux: 4.15.0-87.87 -proposed tracker (LP: #1861165)

  * Bionic update: upstream stable patchset 2020-01-22 (LP: #1860602)
- scsi: lpfc: Fix discovery failures when target device connectivity bounces
- scsi: mpt3sas: Fix clear pending bit in ioctl status
- scsi: lpfc: Fix locking on mailbox command completion
- Input: atmel_mxt_ts - disable IRQ across suspend
- iommu/tegra-smmu: Fix page tables in > 4 GiB memory
- scsi: target: compare full CHAP_A Algorithm strings
- scsi: lpfc: Fix SLI3 hba in loop mode not discovering devices
- scsi: csiostor: Don't enable IRQs too early
- powerpc/pseries: Mark accumulate_stolen_time() as notrace
- powerpc/pseries: Don't fail hash page table insert for bolted mapping
- powerpc/tools: Don't quote $objdump in scripts
- dma-debug: add a schedule point in debug_dma_dump_mappings()
- clocksource/drivers/asm9260: Add a check for of_clk_get
- powerpc/security/book3s64: Report L1TF status in sysfs
- powerpc/book3s64/hash: Add cond_resched to avoid soft lockup warning
- ext4: update direct I/O read lock pattern for IOCB_NOWAIT
- jbd2: Fix statistics for the number of logged blocks
- scsi: tracing: Fix handling of TRANSFER LENGTH == 0 for READ(6) and 
WRITE(6)
- scsi: lpfc: Fix duplicate unreg_rpi error in port offline flow
- f2fs: fix to update dir's i_pino during cross_rename
- clk: qcom: Allow constant ratio freq tables for rcg
- irqchip/irq-bcm7038-l1: Enable parent IRQ if necessary
- irqchip: ingenic: Error out if IRQ domain creation failed
- fs/quota: handle overflows of sysctl fs.quota.* and report as unsigned 
long
- scsi: lpfc: fix: Coverity: lpfc_cmpl_els_rsp(): Null pointer dereferences
- scsi: ufs: fix potential bug which ends in system hang
- powerpc/pseries/cmm: Implement release() function for sysfs device
- powerpc/security: Fix wrong message when RFI Flush is disable
- scsi: atari_scsi: sun3_scsi: Set sg_tablesize to 1 instead of SG_NONE
- clk: pxa: fix one of the pxa RTC clocks
- bcache: at least try to shrink 1 node in bch_mca_scan()
- HID: logitech-hidpp: Silence intermittent get_battery_capacity errors
- libnvdimm/btt: fix variable 'rc' set but not used
- HID: Improve Windows Precision Touchpad detection.
- scsi: pm80xx: Fix for SATA device discovery
- scsi: ufs: Fix error handing during hibern8 enter
- scsi: scsi_debug: num_tgts must be >= 0
- scsi: NCR5380: Add disconnect_mask module parameter
- scsi: iscsi: Don't send data to unbound connection
- scsi: target: iscsi: Wait for all commands to finish before freeing a
  session
- gpio: mpc8xxx: Don't overwrite default irq_set_type callback
- apparmor: fix unsigned len comparison with less than zero
- scripts/kallsyms: fix definitely-lost memory leak
- cdrom: respect device capabilities during opening action
- perf script: Fix brstackinsn for AUXTRACE
- perf regs: Make perf_reg_name() return "unknown" instead of NULL
- s390/zcrypt: handle new reply code FILTERED_BY_HYPERVISOR
- libfdt: define INT32_MAX and UINT32_MAX in libfdt_env.h
- s390/cpum_sf: Check for SDBT and SDB consistency
- ocfs2: fix passing zero to 'PTR_ERR' warning
- kernel: sysctl: make drop_caches write-only
- userfaultfd: require CAP_SYS_PTRACE for UFFD_FEATURE_EVENT_FORK
- x86/mce: Fix possibly incorrect severity calculation on AMD
- net, sysctl: Fix compiler warning when only cBPF is present
- netfilter: nf_queue: enqueue skbs with NULL dst
- ALSA: hda - Downgrade error message for single-cmd fallback
- bonding: fix active-backup transition after link failure
- perf strbuf: Remove redundant va_end() in strbuf_addv()
- Make filldir[64]() verify the directory entry filename is valid
- filldir[64]: remove WARN_ON_ONCE() for bad directory entries
- netfilter: ebtables: compat: reject all padding in matches/watchers
- 6pack,mkiss: fix possible deadlock
- netfilter: bridge: make sure to pull arp header in br_nf_forward_arp()
- inetpeer: fix data-race in inet_putpeer / inet_putpeer
- net: add a READ_ONCE() in skb_peek_tail()
- net: icmp: fix data-race in cmp_global_allow()
- hrtimer: Annotate lockless access to timer->state
- spi: fsl: don't map irq during probe
- tty/serial: atmel: fix out of range clock divider handling
- pinctrl: baytrail: 

[Bug 1850540] Re: multi-zone raid0 corruption

2020-02-17 Thread Launchpad Bug Tracker
This bug was fixed in the package linux - 5.3.0-40.32

---
linux (5.3.0-40.32) eoan; urgency=medium

  * eoan/linux: 5.3.0-40.32 -proposed tracker (LP: #1861214)

  * No sof soundcard for 'ASoC: CODEC DAI intel-hdmi-hifi1 not registered' after
modprobe sof (LP: #1860248)
- ASoC: SOF: Intel: fix HDA codec driver probe with multiple controllers

  * ocfs2-tools is causing kernel panics in Ubuntu Focal (Ubuntu-5.4.0-9.12)
(LP: #1852122)
- ocfs2: fix the crash due to call ocfs2_get_dlm_debug once less

  * QAT drivers for C3XXX and C62X not included as modules (LP: #1845959)
- [Config] CRYPTO_DEV_QAT_C3XXX=m, CRYPTO_DEV_QAT_C62X=m and
  CRYPTO_DEV_QAT_DH895xCC=m

  * Eoan update: upstream stable patchset 2020-01-24 (LP: #1860816)
- scsi: lpfc: Fix discovery failures when target device connectivity bounces
- scsi: mpt3sas: Fix clear pending bit in ioctl status
- scsi: lpfc: Fix locking on mailbox command completion
- Input: atmel_mxt_ts - disable IRQ across suspend
- f2fs: fix to update time in lazytime mode
- iommu: rockchip: Free domain on .domain_free
- iommu/tegra-smmu: Fix page tables in > 4 GiB memory
- dmaengine: xilinx_dma: Clear desc_pendingcount in xilinx_dma_reset
- scsi: target: compare full CHAP_A Algorithm strings
- scsi: lpfc: Fix SLI3 hba in loop mode not discovering devices
- scsi: csiostor: Don't enable IRQs too early
- scsi: hisi_sas: Replace in_softirq() check in hisi_sas_task_exec()
- powerpc/pseries: Mark accumulate_stolen_time() as notrace
- powerpc/pseries: Don't fail hash page table insert for bolted mapping
- powerpc/tools: Don't quote $objdump in scripts
- dma-debug: add a schedule point in debug_dma_dump_mappings()
- leds: lm3692x: Handle failure to probe the regulator
- clocksource/drivers/asm9260: Add a check for of_clk_get
- clocksource/drivers/timer-of: Use unique device name instead of timer
- powerpc/security/book3s64: Report L1TF status in sysfs
- powerpc/book3s64/hash: Add cond_resched to avoid soft lockup warning
- ext4: update direct I/O read lock pattern for IOCB_NOWAIT
- ext4: iomap that extends beyond EOF should be marked dirty
- jbd2: Fix statistics for the number of logged blocks
- scsi: tracing: Fix handling of TRANSFER LENGTH == 0 for READ(6) and 
WRITE(6)
- scsi: lpfc: Fix duplicate unreg_rpi error in port offline flow
- f2fs: fix to update dir's i_pino during cross_rename
- clk: qcom: Allow constant ratio freq tables for rcg
- clk: clk-gpio: propagate rate change to parent
- irqchip/irq-bcm7038-l1: Enable parent IRQ if necessary
- irqchip: ingenic: Error out if IRQ domain creation failed
- fs/quota: handle overflows of sysctl fs.quota.* and report as unsigned 
long
- scsi: lpfc: fix: Coverity: lpfc_cmpl_els_rsp(): Null pointer dereferences
- PCI: rpaphp: Fix up pointer to first drc-info entry
- scsi: ufs: fix potential bug which ends in system hang
- powerpc/pseries/cmm: Implement release() function for sysfs device
- PCI: rpaphp: Don't rely on firmware feature to imply drc-info support
- PCI: rpaphp: Annotate and correctly byte swap DRC properties
- PCI: rpaphp: Correctly match ibm, my-drc-index to drc-name when using drc-
  info
- powerpc/security: Fix wrong message when RFI Flush is disable
- scsi: atari_scsi: sun3_scsi: Set sg_tablesize to 1 instead of SG_NONE
- clk: pxa: fix one of the pxa RTC clocks
- bcache: at least try to shrink 1 node in bch_mca_scan()
- HID: quirks: Add quirk for HP MSU1465 PIXART OEM mouse
- HID: logitech-hidpp: Silence intermittent get_battery_capacity errors
- ARM: 8937/1: spectre-v2: remove Brahma-B53 from hardening
- libnvdimm/btt: fix variable 'rc' set but not used
- HID: Improve Windows Precision Touchpad detection.
- HID: rmi: Check that the RMI_STARTED bit is set before unregistering the 
RMI
  transport device
- watchdog: Fix the race between the release of watchdog_core_data and cdev
- scsi: pm80xx: Fix for SATA device discovery
- scsi: ufs: Fix error handing during hibern8 enter
- scsi: scsi_debug: num_tgts must be >= 0
- scsi: NCR5380: Add disconnect_mask module parameter
- scsi: iscsi: Don't send data to unbound connection
- scsi: target: iscsi: Wait for all commands to finish before freeing a
  session
- gpio: mpc8xxx: Don't overwrite default irq_set_type callback
- apparmor: fix unsigned len comparison with less than zero
- scripts/kallsyms: fix definitely-lost memory leak
- powerpc: Don't add -mabi= flags when building with Clang
- cdrom: respect device capabilities during opening action
- perf script: Fix brstackinsn for AUXTRACE
- perf regs: Make perf_reg_name() return "unknown" instead of NULL
- s390/zcrypt: handle new reply code FILTERED_BY_HYPERVISOR
- libfdt: define INT32_MAX and UINT32_MAX in libfdt_env.h
- 

[Bug 1850540] Re: multi-zone raid0 corruption

2020-02-10 Thread dann frazier
= linux/eoan verification =
I created a multi-zone array on current eoan, then upgraded to the -proposed 
kernel:

[4.021562] md/raid0:md0: !!! DEFAULTING TO ALTERNATE LAYOUT !!!
[4.021563] md/raid0: Please set raid0.default_layout to 1 or 2
[4.021564] md/raid0: Read the following page for more information:
[4.021564] md/raid0: https://wiki.ubuntu.com/Kernel/Raid0LayoutMigration

I verified that the array is active.

= linux/bionic verification =
Similarly:
[1.830003] md/raid0:md0: !!! DEFAULTING TO ALTERNATE LAYOUT !!!
[1.835050] md/raid0: Please set raid0.default_layout to 1 or 2
[1.839513] md/raid0: Read the following page for more information:
[1.844259] md/raid0: https://wiki.ubuntu.com/Kernel/Raid0LayoutMigration

= linux/xenial verification =
Similarly:
[3.157653] md: raid0 personality registered for level 0
[3.170468] md/raid0:md0: !!! DEFAULTING TO ALTERNATE LAYOUT !!!
[3.176180] md/raid0: Please set raid0.default_layout to 1 or 2
[3.181684] md/raid0: Read the following page for more information:
[3.187394] md/raid0: https://wiki.ubuntu.com/Kernel/Raid0LayoutMigration

** Tags removed: verification-needed verification-needed-bionic 
verification-needed-disco verification-needed-eoan
** Tags added: verification-done verification-done-bionic 
verification-done-disco verification-done-eoan

** Tags removed: verification-needed-xenial
** Tags added: verification-done-xenial

-- 
You received this bug notification because you are a member of Ubuntu
Bugs, which is subscribed to Ubuntu.
https://bugs.launchpad.net/bugs/1850540

Title:
  multi-zone raid0 corruption

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu-release-notes/+bug/1850540/+subscriptions

-- 
ubuntu-bugs mailing list
ubuntu-bugs@lists.ubuntu.com
https://lists.ubuntu.com/mailman/listinfo/ubuntu-bugs

[Bug 1850540] Re: multi-zone raid0 corruption

2020-02-10 Thread dann frazier
** Tags removed: verification-done verification-done-bionic 
verification-done-disco verification-done-eoan
** Tags added: verification-needed verification-needed-bionic 
verification-needed-disco verification-needed-eoan

** Tags removed: block-proposed-bionic block-proposed-disco block-
proposed-eoan

-- 
You received this bug notification because you are a member of Ubuntu
Bugs, which is subscribed to Ubuntu.
https://bugs.launchpad.net/bugs/1850540

Title:
  multi-zone raid0 corruption

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu-release-notes/+bug/1850540/+subscriptions

-- 
ubuntu-bugs mailing list
ubuntu-bugs@lists.ubuntu.com
https://lists.ubuntu.com/mailman/listinfo/ubuntu-bugs

[Bug 1850540] Re: multi-zone raid0 corruption

2020-02-03 Thread Launchpad Bug Tracker
This bug was fixed in the package mdadm - 4.1~rc1-3~ubuntu18.04.4

---
mdadm (4.1~rc1-3~ubuntu18.04.4) bionic; urgency=medium

  * Introduce "broken" state for RAID0/Linear in mdadm (LP: #1847924)
- d/p/lp1847924-Introduce-new-array-state-broken-for-raid0.patch

  * Disable patches from proposed version 4.1~rc1-3~ubuntu18.04.3 due to
issues with kernel

 -- gpicc...@canonical.com (Guilherme G. Piccoli)  Tue, 14 Jan 2020
16:10:59 -0300

** Changed in: mdadm (Ubuntu Bionic)
   Status: Fix Committed => Fix Released

-- 
You received this bug notification because you are a member of Ubuntu
Bugs, which is subscribed to Ubuntu.
https://bugs.launchpad.net/bugs/1850540

Title:
  multi-zone raid0 corruption

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu-release-notes/+bug/1850540/+subscriptions

-- 
ubuntu-bugs mailing list
ubuntu-bugs@lists.ubuntu.com
https://lists.ubuntu.com/mailman/listinfo/ubuntu-bugs

[Bug 1850540] Re: multi-zone raid0 corruption

2020-01-30 Thread Ubuntu Kernel Bot
This bug is awaiting verification that the kernel in -proposed solves
the problem. Please test the kernel and update this bug with the
results. If the problem is solved, change the tag 'verification-needed-
xenial' to 'verification-done-xenial'. If the problem still exists,
change the tag 'verification-needed-xenial' to 'verification-failed-
xenial'.

If verification is not done by 5 working days from today, this fix will
be dropped from the source code, and this bug will be closed.

See https://wiki.ubuntu.com/Testing/EnableProposed for documentation how
to enable and use -proposed. Thank you!


** Tags added: verification-needed-xenial

-- 
You received this bug notification because you are a member of Ubuntu
Bugs, which is subscribed to Ubuntu.
https://bugs.launchpad.net/bugs/1850540

Title:
  multi-zone raid0 corruption

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu-release-notes/+bug/1850540/+subscriptions

-- 
ubuntu-bugs mailing list
ubuntu-bugs@lists.ubuntu.com
https://lists.ubuntu.com/mailman/listinfo/ubuntu-bugs

[Bug 1850540] Re: multi-zone raid0 corruption

2020-01-28 Thread Khaled El Mously
** Changed in: linux (Ubuntu Bionic)
   Status: In Progress => Fix Committed

** Changed in: linux (Ubuntu Disco)
   Status: In Progress => Fix Committed

** Changed in: linux (Ubuntu Eoan)
   Status: In Progress => Fix Committed

** Changed in: linux (Ubuntu Xenial)
   Status: Confirmed => Fix Committed

-- 
You received this bug notification because you are a member of Ubuntu
Bugs, which is subscribed to Ubuntu.
https://bugs.launchpad.net/bugs/1850540

Title:
  multi-zone raid0 corruption

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu-release-notes/+bug/1850540/+subscriptions

-- 
ubuntu-bugs mailing list
ubuntu-bugs@lists.ubuntu.com
https://lists.ubuntu.com/mailman/listinfo/ubuntu-bugs

[Bug 1850540] Re: multi-zone raid0 corruption

2020-01-23 Thread Connor Kuehl
** Changed in: linux (Ubuntu Trusty)
   Status: Confirmed => Fix Committed

-- 
You received this bug notification because you are a member of Ubuntu
Bugs, which is subscribed to Ubuntu.
https://bugs.launchpad.net/bugs/1850540

Title:
  multi-zone raid0 corruption

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu-release-notes/+bug/1850540/+subscriptions

-- 
ubuntu-bugs mailing list
ubuntu-bugs@lists.ubuntu.com
https://lists.ubuntu.com/mailman/listinfo/ubuntu-bugs

[Bug 1850540] Re: multi-zone raid0 corruption

2020-01-14 Thread Kleber Sacilotto de Souza
The following patches sent on 2019/12/18 have been removed from the
Eoan/Disco/Bionic kernels:

https://lists.ubuntu.com/archives/kernel-team/2019-December/106461.html

Therefore I'm setting the linux tasks back to 'In Progress'.

** Changed in: linux (Ubuntu Bionic)
   Status: Fix Committed => In Progress

** Changed in: linux (Ubuntu Disco)
   Status: Fix Committed => In Progress

** Changed in: linux (Ubuntu Eoan)
   Status: Fix Committed => In Progress

-- 
You received this bug notification because you are a member of Ubuntu
Bugs, which is subscribed to Ubuntu.
https://bugs.launchpad.net/bugs/1850540

Title:
  multi-zone raid0 corruption

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu-release-notes/+bug/1850540/+subscriptions

-- 
ubuntu-bugs mailing list
ubuntu-bugs@lists.ubuntu.com
https://lists.ubuntu.com/mailman/listinfo/ubuntu-bugs

[Bug 1850540] Re: multi-zone raid0 corruption

2020-01-07 Thread Kleber Sacilotto de Souza
** Changed in: linux (Ubuntu Bionic)
   Status: Confirmed => Fix Committed

** Changed in: linux (Ubuntu Disco)
   Status: Confirmed => Fix Committed

** Changed in: linux (Ubuntu Eoan)
   Status: Confirmed => Fix Committed

-- 
You received this bug notification because you are a member of Ubuntu
Bugs, which is subscribed to Ubuntu.
https://bugs.launchpad.net/bugs/1850540

Title:
  multi-zone raid0 corruption

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu-release-notes/+bug/1850540/+subscriptions

-- 
ubuntu-bugs mailing list
ubuntu-bugs@lists.ubuntu.com
https://lists.ubuntu.com/mailman/listinfo/ubuntu-bugs

[Bug 1850540] Re: multi-zone raid0 corruption

2019-12-12 Thread dann frazier
Adding the block-proposed-* tags as per comment #11.

** Tags removed: verification-needed verification-needed-bionic 
verification-needed-disco verification-needed-eoan
** Tags added: block-proposed-bionic block-proposed-disco block-proposed-eoan 
verification-done verification-done-bionic verification-done-disco 
verification-done-eoan

-- 
You received this bug notification because you are a member of Ubuntu
Bugs, which is subscribed to Ubuntu.
https://bugs.launchpad.net/bugs/1850540

Title:
  multi-zone raid0 corruption

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu-release-notes/+bug/1850540/+subscriptions

-- 
ubuntu-bugs mailing list
ubuntu-bugs@lists.ubuntu.com
https://lists.ubuntu.com/mailman/listinfo/ubuntu-bugs

[Bug 1850540] Re: multi-zone raid0 corruption

2019-12-12 Thread dann frazier
There is no plan to reapply the kernel side of these fixes ahead of
18.04.4. Introducing the mdadm side of that in 18.04.4 therefore would
cause a regression in functionality:

ubuntu@ip-172-30-0-208:~$ sudo mdadm --create /dev/md0 --run
--metadata=default --level=0 --raid-devices=2 /dev/xvdb /dev/xvdc
mdadm: /dev/xvdb appears to be part of a raid array:
   level=raid0 devices=2 ctime=Wed Dec 11 22:10:59 2019
mdadm: /dev/xvdc appears to be part of a raid array:
   level=raid0 devices=2 ctime=Wed Dec 11 22:10:59 2019
mdadm: ADD_NEW_DISK for /dev/xvdb failed: Invalid argument
mdadm: Possibly your kernel doesn't support RAID0 layouts.
mdadm: Either upgrade, or use --layout=dangerous

Of course, at install time, upgrading the kernel isn't an option. So
let's hold this SRU until after the point release.

** Description changed:

  Bug 1849682 tracks the temporarily revert of the fix for this issue,
  while this bug tracks the re-application of that fix once we have a full
  solution.
  
  [Impact]
  (cut & paste from https://marc.info/?l=linux-raid=157360088014027=2)
  An unintentional RAID0 layout change was introduced in the v3.14 kernel. This 
effectively means there are 2 different layouts Linux will use to write data to 
RAID0 arrays in the wild - the “pre-3.14” way and the “3.14 and later” way. 
Mixing these layouts by writing to an array while booted on these different 
kernel versions can lead to corruption.
  
  Note that this only impacts RAID0 arrays that include devices of
  different sizes. If your devices are all the same size, both layouts are
  equivalent, and your array is not at risk of corruption due to this
  issue.
  
  Unfortunately, the kernel cannot detect which layout was used for writes
  to pre-existing arrays, and therefore requires input from the
  administrator. This input can be provided via the kernel command line
  with the raid0.default_layout= parameter, or by setting the
  default_layout module parameter when loading the raid0 module. With a
  new enough version of mdadm (>= 4.2, or equivalent distro backports),
  you can set the layout version when assembling a stopped array. For
  example:
  
  mdadm --stop /dev/md0
  mdadm --assemble -U layout-alternate /dev/md0 /dev/sda1 /dev/sda2
  See the mdadm manpage for more details. Once set in this manner, the layout 
will be recorded in the array and will not need to be explicitly specified in 
the future.
  
  (The mdadm part of this SRU is for the above support ^)
  
  [Test Case]
  = mdadm =
  Confirm that a multi-zone raid0 created w/ older mdadm is able to be started 
on a fixed kernel by setting a layout.
  1) Ex: w/ old kernel/mdadm:
    mdadm --create /dev/md0 --run --metadata=default \
  --level=0 --raid-devices=2 /dev/vdb1 /dev/vdc1
  2) Reboot onto fixed kernel & update mdadm
  3) sudo mdadm --stop /dev/md0 &&
-sudo mdadm --assemble -U layout-alternate \
+    sudo mdadm --assemble -U layout-alternate \
   /dev/md0 /dev/vdb1 /dev/vdc1
  4) Confirm that the array autostarts on reboot
  5) Confirm that w/ new kernel & new mdadm, a user can create and start an 
array in a backwards-compatible fashion (i.e. w/o an explicit layout).
  6) Verify that 'mdadm --detail /dev/md0' displays the layout
  
  = linux =
  Similar to above, but using kernel command line options.
  
  [Regression Risk]
  The kernel side of things will break starting pre-existing arrays. That's 
intentional.
  
- Although I've done due-diligence to check for backwards compatibility
- issues, the mdadm side may still present some.
+ The mdadm side will cause a regression in functionality where a user can
+ no longer create multi-zone raid0s on kernels that do not yet have the
+ raid0 layout patches. This is intentional, as such RAID arrays present a
+ corruption risk.

** Tags removed: verification-done verification-done-bionic 
verification-done-disco verification-done-eoan
** Tags added: verification-needed verification-needed-bionic 
verification-needed-disco verification-needed-eoan

-- 
You received this bug notification because you are a member of Ubuntu
Bugs, which is subscribed to Ubuntu.
https://bugs.launchpad.net/bugs/1850540

Title:
  multi-zone raid0 corruption

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu-release-notes/+bug/1850540/+subscriptions

-- 
ubuntu-bugs mailing list
ubuntu-bugs@lists.ubuntu.com
https://lists.ubuntu.com/mailman/listinfo/ubuntu-bugs

[Bug 1850540] Re: multi-zone raid0 corruption

2019-12-06 Thread dann frazier
= bionic verification =
ubuntu@ip-172-30-0-117:~$ sudo mdadm --create /dev/md0 --run --metadata=default 
--level=0 --raid-devices=4 /dev/xvdb /dev/xvdc /dev/xvdd /dev/xvde
mdadm: array /dev/md0 started.
ubuntu@ip-172-30-0-117:~$ cat /proc/mdstat
Personalities : [linear] [multipath] [raid0] [raid1] [raid6] [raid5] [raid4] 
[raid10] 
md0 : active raid0 xvde[3] xvdd[2] xvdc[1] xvdb[0]
  29323264 blocks super 1.2 512k chunks
  
unused devices: 

## kernel & mdadm upgrade

ubuntu@ip-172-30-0-117:~$ dmesg | grep raid
[4.086107] md/raid0:md0: cannot assemble multi-zone RAID0 with 
default_layout setting
[4.092253] md/raid0: please set raid0.default_layout to 1 or 2
[4.452725] raid6: avx2x4   gen() 24200 MB/s
[4.500724] raid6: avx2x4   xor() 15657 MB/s
[4.548725] raid6: avx2x2   gen() 21010 MB/s
[4.596724] raid6: avx2x2   xor() 13248 MB/s
[4.644724] raid6: avx2x1   gen() 18005 MB/s
[4.692726] raid6: avx2x1   xor() 12606 MB/s
[4.740726] raid6: sse2x4   gen() 13375 MB/s
[4.788725] raid6: sse2x4   xor()  8309 MB/s
[4.836730] raid6: sse2x2   gen() 11042 MB/s
[4.884725] raid6: sse2x2   xor()  7264 MB/s
[4.932730] raid6: sse2x1   gen()  9288 MB/s
[4.980723] raid6: sse2x1   xor()  6578 MB/s
[4.983827] raid6: using algorithm avx2x4 gen() 24200 MB/s
[4.987535] raid6:  xor() 15657 MB/s, rmw enabled
[4.991018] raid6: using avx2x2 recovery algorithm
ubuntu@ip-172-30-0-117:~$ cat /proc/mdstat
Personalities : [raid0] [linear] [multipath] [raid1] [raid6] [raid5] [raid4] 
[raid10] 
md0 : inactive xvdd[2] xvde[3] xvdb[0] xvdc[1]
  29323264 blocks super 1.2
   
unused devices: 
ubuntu@ip-172-30-0-117:~$ sudo mdadm --stop /dev/md0
mdadm: stopped /dev/md0
ubuntu@ip-172-30-0-117:~$ sudo mdadm --assemble /dev/md0 -U layout-alternate 
/dev/xvdb /dev/xvdc /dev/xvdd /dev/xvde
mdadm: /dev/md0 has been started with 4 drives.
ubuntu@ip-172-30-0-117:~$ cat /proc/mdstat
Personalities : [raid0] [linear] [multipath] [raid1] [raid6] [raid5] [raid4] 
[raid10] 
md0 : active raid0 xvdb[0] xvde[3] xvdd[2] xvdc[1]
  29323264 blocks super 1.2 512k chunks
  
unused devices: 

## reboot
ubuntu@ip-172-30-0-117:~$ dmesg | grep raid
[3.793154] raid6: avx2x4   gen() 24292 MB/s
[3.841155] raid6: avx2x4   xor() 15646 MB/s
[3.889157] raid6: avx2x2   gen() 20570 MB/s
[3.937155] raid6: avx2x2   xor() 13351 MB/s
[3.985155] raid6: avx2x1   gen() 18190 MB/s
[4.033154] raid6: avx2x1   xor() 12469 MB/s
[4.081153] raid6: sse2x4   gen() 13399 MB/s
[4.129153] raid6: sse2x4   xor()  8358 MB/s
[4.177151] raid6: sse2x2   gen() 10984 MB/s
[4.225157] raid6: sse2x2   xor()  7224 MB/s
[4.273158] raid6: sse2x1   gen()  9335 MB/s
[4.321157] raid6: sse2x1   xor()  6578 MB/s
[4.323567] raid6: using algorithm avx2x4 gen() 24292 MB/s
[4.326583] raid6:  xor() 15646 MB/s, rmw enabled
[4.329350] raid6: using avx2x2 recovery algorithm
ubuntu@ip-172-30-0-117:~$ cat /proc/mdstat
Personalities : [raid0] [linear] [multipath] [raid1] [raid6] [raid5] [raid4] 
[raid10] 
md0 : active raid0 xvde[3] xvdb[0] xvdc[1] xvdd[2]
  29323264 blocks super 1.2 512k chunks
  
unused devices: 
ubuntu@ip-172-30-0-117:~$ sudo mdadm --stop /dev/md0
mdadm: stopped /dev/md0

ubuntu@ip-172-30-0-117:~$ sudo mdadm --create /dev/md0 --run --metadata=default 
--level=0 --raid-devices=3 /dev/xvdb /dev/xvdd /dev/xvde
mdadm: /dev/xvdb appears to be part of a raid array:
   level=raid0 devices=4 ctime=Fri Dec  6 22:20:04 2019
mdadm: /dev/xvdd appears to be part of a raid array:
   level=raid0 devices=4 ctime=Fri Dec  6 22:20:04 2019
mdadm: /dev/xvde appears to be part of a raid array:
   level=raid0 devices=4 ctime=Fri Dec  6 22:20:04 2019
mdadm: array /dev/md0 started.
ubuntu@ip-172-30-0-117:~$ sudo mdadm --detail /dev/md0 | grep Layout
Layout : original


** Tags removed: verification-needed verification-needed-bionic
** Tags added: verification-done verification-done-bionic

-- 
You received this bug notification because you are a member of Ubuntu
Bugs, which is subscribed to Ubuntu.
https://bugs.launchpad.net/bugs/1850540

Title:
  multi-zone raid0 corruption

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu-release-notes/+bug/1850540/+subscriptions

-- 
ubuntu-bugs mailing list
ubuntu-bugs@lists.ubuntu.com
https://lists.ubuntu.com/mailman/listinfo/ubuntu-bugs

[Bug 1850540] Re: multi-zone raid0 corruption

2019-12-06 Thread dann frazier
** Tags removed: verification-needed-disco
** Tags added: verification-done-disco

-- 
You received this bug notification because you are a member of Ubuntu
Bugs, which is subscribed to Ubuntu.
https://bugs.launchpad.net/bugs/1850540

Title:
  multi-zone raid0 corruption

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu-release-notes/+bug/1850540/+subscriptions

-- 
ubuntu-bugs mailing list
ubuntu-bugs@lists.ubuntu.com
https://lists.ubuntu.com/mailman/listinfo/ubuntu-bugs

[Bug 1850540] Re: multi-zone raid0 corruption

2019-12-06 Thread dann frazier
= disco verification =
ubuntu@ip-172-30-0-118:~$ sudo mdadm --create /dev/md0 --run --metadata=default 
--level=0 --raid-devices=2 /dev/xvdb /dev/xvdc
mdadm: Fail create md0 when using /sys/module/md_mod/parameters/new_array
mdadm: array /dev/md0 started.
ubuntu@ip-172-30-0-118:~$ cat /proc/mdstat
Personalities : [raid0] 
md0 : active raid0 xvdc[1] xvdb[0]
  3141632 blocks super 1.2 512k chunks
  
unused devices: 

## kernel & mdadm upgrade
ubuntu@ip-172-30-0-118:~$ dmesg | grep raid
[2.419732] md: If you don't use raid, use raid=noautodetect
[4.990204] md/raid0:md0: cannot assemble multi-zone RAID0 with 
default_layout setting
[4.997283] md/raid0: please set raid0.default_layout to 1 or 2
ubuntu@ip-172-30-0-118:~$ sudo mdadm --stop /dev/md0
mdadm: stopped /dev/md0
ubuntu@ip-172-30-0-118:~$ sudo mdadm --assemble -U layout-original /dev/md0 
/dev/xvdb /dev/xvdc
mdadm: /dev/md0 has been started with 2 drives.
ubuntu@ip-172-30-0-118:~$ cat /proc/mdstat
Personalities : [raid0] 
md0 : active raid0 xvdb[0] xvdc[1]
  3141632 blocks super 1.2 512k chunks
  
unused devices: 

## reboot

ubuntu@ip-172-30-0-118:~$ dmesg | grep raid
[2.545479] md: If you don't use raid, use raid=noautodetect
ubuntu@ip-172-30-0-118:~$ cat /proc/mdstat
Personalities : [raid0] 
md0 : active raid0 xvdb[0] xvdc[1]
  3141632 blocks super 1.2 512k chunks
  
unused devices: 
ubuntu@ip-172-30-0-118:~$ sudo mdadm --stop /dev/md0
mdadm: stopped /dev/md0
ubuntu@ip-172-30-0-118:~$ sudo mdadm --create /dev/md0 --run --metadata=default 
--level=0 --raid-devices=2 /dev/xvdc /dev/xvdb
mdadm: /dev/xvdc appears to be part of a raid array:
   level=raid0 devices=2 ctime=Fri Dec  6 21:57:50 2019
mdadm: /dev/xvdb appears to be part of a raid array:
   level=raid0 devices=2 ctime=Fri Dec  6 21:57:50 2019
mdadm: array /dev/md0 started.
ubuntu@ip-172-30-0-118:~$ sudo mdadm --detail /dev/md0 | grep Layout
Layout : original

-- 
You received this bug notification because you are a member of Ubuntu
Bugs, which is subscribed to Ubuntu.
https://bugs.launchpad.net/bugs/1850540

Title:
  multi-zone raid0 corruption

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu-release-notes/+bug/1850540/+subscriptions

-- 
ubuntu-bugs mailing list
ubuntu-bugs@lists.ubuntu.com
https://lists.ubuntu.com/mailman/listinfo/ubuntu-bugs

[Bug 1850540] Re: multi-zone raid0 corruption

2019-12-06 Thread dann frazier
= eoan verification =

ubuntu@ip-172-30-0-147:~$ sudo mdadm --create /dev/md0 --run --metadata=default 
--level=0 --raid-devices=2 /dev/xvdb /dev/xvdc 
mdadm: Fail create md0 when using /sys/module/md_mod/parameters/new_array
mdadm: array /dev/md0 started.
ubuntu@ip-172-30-0-147:~$ cat /proc/mdstat
Personalities : [raid0] 
md0 : active raid0 xvdc[1] xvdb[0]
  3141632 blocks super 1.2 512k chunks
  
unused devices: 

# Upgraded mdadm & installed a 5.4 kernel
ubuntu@ip-172-30-0-147:~$ dmesg | grep raid
[3.009968] md: If you don't use raid, use raid=noautodetect
[5.981265] md/raid0:md0: cannot assemble multi-zone RAID0 with 
default_layout setting
[5.993787] md/raid0: please set raid0.default_layout to 1 or 2
ubuntu@ip-172-30-0-147:~$ sudo mdadm --stop /dev/md0
mdadm: stopped /dev/md0
ubuntu@ip-172-30-0-147:~$ sudo mdadm --assemble -U layout-alternate /dev/md0 
/dev/xvdb /dev/xvdc
mdadm: /dev/md0 has been started with 2 drives.
ubuntu@ip-172-30-0-147:~$ cat /proc/mdstat
Personalities : [raid0] 
md0 : active raid0 xvdb[0] xvdc[1]
  3141632 blocks super 1.2 512k chunks
  
unused devices: 

## reboot
ubuntu@ip-172-30-0-147:~$ dmesg | grep raid
[2.440539] md: If you don't use raid, use raid=noautodetect
ubuntu@ip-172-30-0-147:~$ cat /proc/mdstat
Personalities : [raid0] 
md0 : active raid0 xvdb[0] xvdc[1]
  3141632 blocks super 1.2 512k chunks
  
unused devices: 
ubuntu@ip-172-30-0-147:~$ sudo mdadm --stop /dev/md0
mdadm: stopped /dev/md0
ubuntu@ip-172-30-0-147:~$ sudo mdadm --create /dev/md0 --run --metadata=default 
--level=0 --raid-devices=2 /dev/xvdc /dev/xvdb
mdadm: /dev/xvdc appears to be part of a raid array:
   level=raid0 devices=2 ctime=Fri Dec  6 21:30:03 2019
mdadm: /dev/xvdb appears to be part of a raid array:
   level=raid0 devices=2 ctime=Fri Dec  6 21:30:03 2019
mdadm: array /dev/md0 started.
ubuntu@ip-172-30-0-147:~$ sudo mdadm --stop /dev/md0
mdadm: stopped /dev/md0
ubuntu@ip-172-30-0-147:~$ sudo mdadm --create /dev/md0 --run --metadata=default 
--level=0 --raid-devices=2 /dev/xvdc /dev/xvdb --layout=original
mdadm: /dev/xvdc appears to be part of a raid array:
   level=raid0 devices=2 ctime=Fri Dec  6 21:52:42 2019
mdadm: /dev/xvdb appears to be part of a raid array:
   level=raid0 devices=2 ctime=Fri Dec  6 21:52:42 2019
mdadm: array /dev/md0 started.
ubuntu@ip-172-30-0-147:~$ sudo mdadm --detail /dev/md0 | grep Layout
Layout : original


** Description changed:

  Bug 1849682 tracks the temporarily revert of the fix for this issue,
  while this bug tracks the re-application of that fix once we have a full
  solution.
  
  [Impact]
  (cut & paste from https://marc.info/?l=linux-raid=157360088014027=2)
  An unintentional RAID0 layout change was introduced in the v3.14 kernel. This 
effectively means there are 2 different layouts Linux will use to write data to 
RAID0 arrays in the wild - the “pre-3.14” way and the “3.14 and later” way. 
Mixing these layouts by writing to an array while booted on these different 
kernel versions can lead to corruption.
  
  Note that this only impacts RAID0 arrays that include devices of
  different sizes. If your devices are all the same size, both layouts are
  equivalent, and your array is not at risk of corruption due to this
  issue.
  
  Unfortunately, the kernel cannot detect which layout was used for writes
  to pre-existing arrays, and therefore requires input from the
  administrator. This input can be provided via the kernel command line
  with the raid0.default_layout= parameter, or by setting the
  default_layout module parameter when loading the raid0 module. With a
  new enough version of mdadm (>= 4.2, or equivalent distro backports),
  you can set the layout version when assembling a stopped array. For
  example:
  
  mdadm --stop /dev/md0
  mdadm --assemble -U layout-alternate /dev/md0 /dev/sda1 /dev/sda2
  See the mdadm manpage for more details. Once set in this manner, the layout 
will be recorded in the array and will not need to be explicitly specified in 
the future.
  
  (The mdadm part of this SRU is for the above support ^)
  
  [Test Case]
  = mdadm =
  Confirm that a multi-zone raid0 created w/ older mdadm is able to be started 
on a fixed kernel by setting a layout.
  1) Ex: w/ old kernel/mdadm:
-   mdadm --create /dev/md0 --run --metadata=default \
- --level=0 --raid-devices=2 /dev/vdb1 /dev/vdc1
+   mdadm --create /dev/md0 --run --metadata=default \
+ --level=0 --raid-devices=2 /dev/vdb1 /dev/vdc1
  2) Reboot onto fixed kernel & update mdadm
- 3) sudo mdadm --assemble -U layout-alternate \
-  /dev/md0 /dev/vdb1 /dev/vdc1
+ 3) sudo mdadm --stop /dev/md0 &&
+sudo mdadm --assemble -U layout-alternate \
+  /dev/md0 /dev/vdb1 /dev/vdc1
  4) Confirm that the array autostarts on reboot
  5) Confirm that w/ new kernel & new mdadm, a user can create and start an 
array in a backwards-compatible fashion (i.e. w/o an explicit layout).
- 6) Verify 

[Bug 1850540] Re: multi-zone raid0 corruption

2019-12-06 Thread Brian Murray
Hello dann, or anyone else affected,

Accepted mdadm into disco-proposed. The package will build now and be
available at https://launchpad.net/ubuntu/+source/mdadm/4.1-1ubuntu1.1
in a few hours, and then in the -proposed repository.

Please help us by testing this new package.  See
https://wiki.ubuntu.com/Testing/EnableProposed for documentation on how
to enable and use -proposed.  Your feedback will aid us getting this
update out to other Ubuntu users.

If this package fixes the bug for you, please add a comment to this bug,
mentioning the version of the package you tested and change the tag from
verification-needed-disco to verification-done-disco. If it does not fix
the bug for you, please add a comment stating that, and change the tag
to verification-failed-disco. In either case, without details of your
testing we will not be able to proceed.

Further information regarding the verification process can be found at
https://wiki.ubuntu.com/QATeam/PerformingSRUVerification .  Thank you in
advance for helping!

N.B. The updated package will be released to -updates after the bug(s)
fixed by this package have been verified and the package has been in
-proposed for a minimum of 7 days.

** Changed in: mdadm (Ubuntu Disco)
   Status: In Progress => Fix Committed

** Tags added: verification-needed-disco

** Changed in: mdadm (Ubuntu Bionic)
   Status: In Progress => Fix Committed

** Tags added: verification-needed-bionic

-- 
You received this bug notification because you are a member of Ubuntu
Bugs, which is subscribed to Ubuntu.
https://bugs.launchpad.net/bugs/1850540

Title:
  multi-zone raid0 corruption

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu-release-notes/+bug/1850540/+subscriptions

-- 
ubuntu-bugs mailing list
ubuntu-bugs@lists.ubuntu.com
https://lists.ubuntu.com/mailman/listinfo/ubuntu-bugs

[Bug 1850540] Re: multi-zone raid0 corruption

2019-12-06 Thread Brian Murray
Hello dann, or anyone else affected,

Accepted mdadm into eoan-proposed. The package will build now and be
available at https://launchpad.net/ubuntu/+source/mdadm/4.1-2ubuntu3.1
in a few hours, and then in the -proposed repository.

Please help us by testing this new package.  See
https://wiki.ubuntu.com/Testing/EnableProposed for documentation on how
to enable and use -proposed.  Your feedback will aid us getting this
update out to other Ubuntu users.

If this package fixes the bug for you, please add a comment to this bug,
mentioning the version of the package you tested and change the tag from
verification-needed-eoan to verification-done-eoan. If it does not fix
the bug for you, please add a comment stating that, and change the tag
to verification-failed-eoan. In either case, without details of your
testing we will not be able to proceed.

Further information regarding the verification process can be found at
https://wiki.ubuntu.com/QATeam/PerformingSRUVerification .  Thank you in
advance for helping!

N.B. The updated package will be released to -updates after the bug(s)
fixed by this package have been verified and the package has been in
-proposed for a minimum of 7 days.

** Changed in: mdadm (Ubuntu Eoan)
   Status: In Progress => Fix Committed

** Tags added: verification-needed verification-needed-eoan

-- 
You received this bug notification because you are a member of Ubuntu
Bugs, which is subscribed to Ubuntu.
https://bugs.launchpad.net/bugs/1850540

Title:
  multi-zone raid0 corruption

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu-release-notes/+bug/1850540/+subscriptions

-- 
ubuntu-bugs mailing list
ubuntu-bugs@lists.ubuntu.com
https://lists.ubuntu.com/mailman/listinfo/ubuntu-bugs

[Bug 1850540] Re: multi-zone raid0 corruption

2019-12-04 Thread dann frazier
** Description changed:

  Bug 1849682 tracks the temporarily revert of the fix for this issue,
  while this bug tracks the re-application of that fix once we have a full
  solution.
  
- Fix checklist:
- [ ] Restore c84a1372df929 md/raid0: avoid RAID0 data corruption due to layout 
confusion.
- [ ] Also apply these fixes:
- 33f2c35a54dfd md: add feature flag MD_FEATURE_RAID0_LAYOUT
- 3874d73e06c9b md/raid0: fix warning message for parameter default_layout
- [ ] If upstream, include https://marc.info/?l=linux-raid=157239231220119=2
- [ ] mdadm update (see Comment #2)
- [ ] Packaging work to detect/aide admin before reboot
+ [Impact]
+ (cut & paste from https://marc.info/?l=linux-raid=157360088014027=2)
+ An unintentional RAID0 layout change was introduced in the v3.14 kernel. This 
effectively means there are 2 different layouts Linux will use to write data to 
RAID0 arrays in the wild - the “pre-3.14” way and the “3.14 and later” way. 
Mixing these layouts by writing to an array while booted on these different 
kernel versions can lead to corruption.
  
- Users of RAID0 arrays are susceptible to a corruption issue if:
-  - The members of the RAID array are not all the same size[*]
-  - Data has been written to the array while running kernels < 3.14 *and* >= 
3.14.
+ Note that this only impacts RAID0 arrays that include devices of
+ different sizes. If your devices are all the same size, both layouts are
+ equivalent, and your array is not at risk of corruption due to this
+ issue.
  
- This is because of an change in v3.14 that accidentally changed how data was 
written - as described in the upstream commit message:
- 
https://github.com/torvalds/linux/commit/c84a1372df929033cb1a0441fb57bd3932f39ac9
+ Unfortunately, the kernel cannot detect which layout was used for writes
+ to pre-existing arrays, and therefore requires input from the
+ administrator. This input can be provided via the kernel command line
+ with the raid0.default_layout= parameter, or by setting the
+ default_layout module parameter when loading the raid0 module. With a
+ new enough version of mdadm (>= 4.2, or equivalent distro backports),
+ you can set the layout version when assembling a stopped array. For
+ example:
  
- That change has been applied to stable, but we reverted it to fix
- 1849682 until we have a full solution ready.
+ mdadm --stop /dev/md0
+ mdadm --assemble -U layout-alternate /dev/md0 /dev/sda1 /dev/sda2
+ See the mdadm manpage for more details. Once set in this manner, the layout 
will be recorded in the array and will not need to be explicitly specified in 
the future.
  
- To summarize, upstream is dealing with this by adding a versioned layout
- in v5.4, and that is being backported to stable kernels - which is why
- we're now seeing it. Layout version 1 is the pre-3.14 layout, version 2
- is post 3.14. Mixing version 1 & version 2 layouts can cause corruption.
- However, until an mdadm exists that is able to set a layout in the
- array, there's no way for the kernel to know which version(s) was used
- to write the existing data. This undefined mode is considered "Version
- 0", and the kernel will now refuse to start these arrays w/o user
- intervention.
+ (The mdadm part of this SRU is for the above support ^)
  
- The user experience is pretty awful here. A user upgrades to the next
- SRU and all of a sudden their system stops at an (initramfs) prompt. A
- clueful user can spot something like the following in dmesg:
+ [Test Case]
+ = mdadm =
+ Confirm that a multi-zone raid0 created w/ older mdadm is able to be started 
on a fixed kernel by setting a layout.
+ 1) Ex: w/ old kernel/mdadm:
+   mdadm --create /dev/md0 --run --metadata=default \
+ --level=0 --raid-devices=2 /dev/vdb1 /dev/vdc1
+ 2) Reboot onto fixed kernel & update mdadm
+ 3) sudo mdadm --assemble -U layout-alternate \
+  /dev/md0 /dev/vdb1 /dev/vdc1
+ 4) Confirm that the array autostarts on reboot
+ 5) Confirm that w/ new kernel & new mdadm, a user can create and start an 
array in a backwards-compatible fashion (i.e. w/o an explicit layout).
+ 6) Verify that 'mdadm --detail /dev/md0' displays the layout 
  
- Here's the message which , as you can see from the log in Comment #1, is
- hidden in a ton of other messages:
+ = linux =
+ Similar to above, but using kernel command line options.
  
- [ 72.720232] md/raid0:md0: cannot assemble multi-zone RAID0 with 
default_layout setting
- [ 72.728149] md/raid0: please set raid.default_layout to 1 or 2
- [ 72.733979] md: pers->run() failed ...
- mdadm: failed to start array /dev/md0: Unknown error 524
+ [Regression Risk]
+ The kernel side of things will break starting pre-existing arrays. That's 
intentional.
  
- What that is trying to say is that you should determine if your data -
- specifically the data toward the end of your array - was most likely
- written with a pre-3.14 or post-3.14 kernel. Based on that, reboot with
- the kernel parameter raid0.default_layout=1 or 

[Bug 1850540] Re: multi-zone raid0 corruption

2019-12-04 Thread Launchpad Bug Tracker
This bug was fixed in the package mdadm - 4.1-4ubuntu1

---
mdadm (4.1-4ubuntu1) focal; urgency=medium

  [ dann frazier ]
  * Merge from Debian unstable.  Remaining changes:
- Ship finalrd hook.
- Do not install mdadm-shutdown.service on Ubuntu.
- Drop broken and unused init scripts in favor of native systemd units,
  which can cause failure to reconfigure mdadm package under certain
  confiment types.
- Drop /etc/cron.d/mdadm and migrate to systemd mdcheck_start|continue
  timer units.
- Drop /etc/cron.daily/mdadm and migrate to system mdmonitor-oneshot
  timer unit.
- mdcheck_start.timer configures the mdcheck on a first sunday of the
  month, with a randomized start delay of up to 24h, and runs for at
  most 6h. mdcheck_continue.timer kicks off daily, with a randomized
  start delay of up to 12h, and continues mdcheck for at most 6h.
- mdmonitor-oneshot.timer runs daily, with a randomized start delay of
  up to 24h.
- One can use systemd drop-ins to change .timer units timings, set
  environmental variables to decrease/increase the length of checking,
  or start the checks by hand. Previously used checkarray is still
  available, albeit not used by timer units.
- Above ensures that previous daily / monthly checks are performed, but
  are randomized, such that performance is not as impacted across a
  cluster of machines.
  * Honor the debconf daily autoscan setting in the systemd timer.

  [ Guilherme G. Piccoli ]
  * Introduce "broken" state for RAID0/Linear in mdadm (LP: #1847924)

 -- dann frazier   Wed, 04 Dec 2019 07:05:07 -0700

** Changed in: mdadm (Ubuntu Focal)
   Status: Confirmed => Fix Released

-- 
You received this bug notification because you are a member of Ubuntu
Bugs, which is subscribed to Ubuntu.
https://bugs.launchpad.net/bugs/1850540

Title:
  multi-zone raid0 corruption

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu-release-notes/+bug/1850540/+subscriptions

-- 
ubuntu-bugs mailing list
ubuntu-bugs@lists.ubuntu.com
https://lists.ubuntu.com/mailman/listinfo/ubuntu-bugs

[Bug 1850540] Re: multi-zone raid0 corruption

2019-12-03 Thread Bug Watch Updater
** Changed in: mdadm (Debian)
   Status: New => Fix Released

-- 
You received this bug notification because you are a member of Ubuntu
Bugs, which is subscribed to Ubuntu.
https://bugs.launchpad.net/bugs/1850540

Title:
  multi-zone raid0 corruption

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu-release-notes/+bug/1850540/+subscriptions

-- 
ubuntu-bugs mailing list
ubuntu-bugs@lists.ubuntu.com
https://lists.ubuntu.com/mailman/listinfo/ubuntu-bugs

[Bug 1850540] Re: multi-zone raid0 corruption

2019-11-13 Thread Bug Watch Updater
** Changed in: mdadm (Debian)
   Status: Unknown => New

-- 
You received this bug notification because you are a member of Ubuntu
Bugs, which is subscribed to Ubuntu.
https://bugs.launchpad.net/bugs/1850540

Title:
  multi-zone raid0 corruption

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu-release-notes/+bug/1850540/+subscriptions

-- 
ubuntu-bugs mailing list
ubuntu-bugs@lists.ubuntu.com
https://lists.ubuntu.com/mailman/listinfo/ubuntu-bugs

[Bug 1850540] Re: multi-zone raid0 corruption

2019-11-13 Thread dann frazier
** Bug watch added: Debian Bug tracker #944676
   https://bugs.debian.org/cgi-bin/bugreport.cgi?bug=944676

** Also affects: mdadm (Debian) via
   https://bugs.debian.org/cgi-bin/bugreport.cgi?bug=944676
   Importance: Unknown
   Status: Unknown

** Also affects: ubuntu-release-notes
   Importance: Undecided
   Status: New

-- 
You received this bug notification because you are a member of Ubuntu
Bugs, which is subscribed to Ubuntu.
https://bugs.launchpad.net/bugs/1850540

Title:
  multi-zone raid0 corruption

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu-release-notes/+bug/1850540/+subscriptions

-- 
ubuntu-bugs mailing list
ubuntu-bugs@lists.ubuntu.com
https://lists.ubuntu.com/mailman/listinfo/ubuntu-bugs

[Bug 1850540] Re: multi-zone raid0 corruption

2019-11-01 Thread dann frazier
mdadm patches are under review upstream:
  https://marc.info/?l=linux-raid=157247979712647=2

** Changed in: mdadm (Ubuntu Focal)
   Status: New => Confirmed

** Changed in: mdadm (Ubuntu Eoan)
   Status: New => Confirmed

** Changed in: mdadm (Ubuntu Disco)
   Status: New => Confirmed

** Changed in: mdadm (Ubuntu Bionic)
   Status: New => Confirmed

** Changed in: mdadm (Ubuntu Xenial)
   Status: New => Confirmed

** Changed in: mdadm (Ubuntu Trusty)
   Status: New => Confirmed

** Changed in: linux (Ubuntu Focal)
   Status: Incomplete => Confirmed

** Changed in: linux (Ubuntu Eoan)
   Status: Incomplete => Confirmed

** Changed in: linux (Ubuntu Disco)
   Status: Incomplete => Confirmed

** Changed in: linux (Ubuntu Bionic)
   Status: Incomplete => Confirmed

** Changed in: linux (Ubuntu Xenial)
   Status: Incomplete => Confirmed

** Changed in: linux (Ubuntu Trusty)
   Status: Incomplete => Confirmed

** Changed in: linux (Ubuntu Precise)
   Status: Incomplete => New

-- 
You received this bug notification because you are a member of Ubuntu
Bugs, which is subscribed to Ubuntu.
https://bugs.launchpad.net/bugs/1850540

Title:
  multi-zone raid0 corruption

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/linux/+bug/1850540/+subscriptions

-- 
ubuntu-bugs mailing list
ubuntu-bugs@lists.ubuntu.com
https://lists.ubuntu.com/mailman/listinfo/ubuntu-bugs

[Bug 1850540] Re: multi-zone raid0 corruption

2019-10-30 Thread dann frazier
** Description changed:

  Bug 1849682 tracks the temporarily revert of the fix for this issue,
  while this bug tracks the re-application of that fix once we have a full
  solution.
+ 
+ Fix checklist:
+ [ ] Restore c84a1372df929 md/raid0: avoid RAID0 data corruption due to layout 
confusion.
+ [ ] Also apply these fixes:
+ 33f2c35a54dfd md: add feature flag MD_FEATURE_RAID0_LAYOUT
+ 3874d73e06c9b md/raid0: fix warning message for parameter default_layout
+ [ ] If upstream, include https://marc.info/?l=linux-raid=157239231220119=2
+ [ ] mdadm update (see Comment #2)
+ [ ] Packaging work to detect/aide admin before reboot
  
  Users of RAID0 arrays are susceptible to a corruption issue if:
   - The members of the RAID array are not all the same size[*]
   - Data has been written to the array while running kernels < 3.14 *and* >= 
3.14.
  
  This is because of an change in v3.14 that accidentally changed how data was 
written - as described in the upstream commit message:
  
https://github.com/torvalds/linux/commit/c84a1372df929033cb1a0441fb57bd3932f39ac9
  
  That change has been applied to stable, but we reverted it to fix
  1849682 until we have a full solution ready.
  
  To summarize, upstream is dealing with this by adding a versioned layout
  in v5.4, and that is being backported to stable kernels - which is why
  we're now seeing it. Layout version 1 is the pre-3.14 layout, version 2
  is post 3.14. Mixing version 1 & version 2 layouts can cause corruption.
  However, until an mdadm exists that is able to set a layout in the
  array, there's no way for the kernel to know which version(s) was used
  to write the existing data. This undefined mode is considered "Version
  0", and the kernel will now refuse to start these arrays w/o user
  intervention.
  
  The user experience is pretty awful here. A user upgrades to the next
  SRU and all of a sudden their system stops at an (initramfs) prompt. A
  clueful user can spot something like the following in dmesg:
  
  Here's the message which , as you can see from the log in Comment #1, is
  hidden in a ton of other messages:
  
  [ 72.720232] md/raid0:md0: cannot assemble multi-zone RAID0 with 
default_layout setting
  [ 72.728149] md/raid0: please set raid.default_layout to 1 or 2
  [ 72.733979] md: pers->run() failed ...
  mdadm: failed to start array /dev/md0: Unknown error 524
  
  What that is trying to say is that you should determine if your data -
  specifically the data toward the end of your array - was most likely
  written with a pre-3.14 or post-3.14 kernel. Based on that, reboot with
  the kernel parameter raid0.default_layout=1 or raid0.default_layout=2 on
  the kernel command line. And note it should be *raid0.default_layout*
  not *raid.default_layout* as the message says - a fix for that message
  is now queued for stable:
  
  
https://github.com/torvalds/linux/commit/3874d73e06c9b9dc15de0b7382fc223986d75571)
  
  IMHO, we should work with upstream to create a web page that clearly
  walks the user through this process, and update the error message to
  point to that page. I'd also like to see if we can detect this problem
  *before* the user reboots (debconf?) and help the user fix things. e.g.
  "We detected that you have RAID0 arrays that maybe susceptible to a
  corruption problem", guide the user to choosing a layout, and update the
  mdadm initramfs hook to poke the answer in via sysfs before starting the
  array on reboot.
  
  Note that it also seems like we should investigate backporting this to <
  3.14 kernels. Imagine a user switching between the trusty HWE kernel and
  the GA kernel.
  
  References from users of other distros:
  https://blog.icod.de/2019/10/10/caution-kernel-5-3-4-and-raid0-default_layout/
  
https://www.linuxquestions.org/questions/linux-general-1/raid-arrays-not-assembling-4175662774/
  
  [*] Which surprisingly is not the case reported in this bug - the user
  here had a raid0 of 8 identically-sized devices. I suspect there's a bug
  in the detection code somewhere.

-- 
You received this bug notification because you are a member of Ubuntu
Bugs, which is subscribed to Ubuntu.
https://bugs.launchpad.net/bugs/1850540

Title:
  multi-zone raid0 corruption

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/linux/+bug/1850540/+subscriptions

-- 
ubuntu-bugs mailing list
ubuntu-bugs@lists.ubuntu.com
https://lists.ubuntu.com/mailman/listinfo/ubuntu-bugs

[Bug 1850540] Re: multi-zone raid0 corruption

2019-10-30 Thread dann frazier
** Description changed:

  Bug 1849682 tracks the temporarily revert of the fix for this issue,
  while this bug tracks the re-application of that fix once we have a full
  solution.
  
  Users of RAID0 arrays are susceptible to a corruption issue if:
   - The members of the RAID array are not all the same size[*]
   - Data has been written to the array while running kernels < 3.14 *and* >= 
3.14.
  
  This is because of an change in v3.14 that accidentally changed how data was 
written - as described in the upstream commit message:
  
https://github.com/torvalds/linux/commit/c84a1372df929033cb1a0441fb57bd3932f39ac9
  
  That change has been applied to stable, but we reverted it to fix
  1849682 until we have a full solution ready.
  
  To summarize, upstream is dealing with this by adding a versioned layout
  in v5.4, and that is being backported to stable kernels - which is why
  we're now seeing it. Layout version 1 is the pre-3.14 layout, version 2
  is post 3.14. Mixing version 1 & version 2 layouts can cause corruption.
- However, unless a layout-version-aware kernel *created* the array,
- there's no way for the kernel to know which version(s) was used to write
- the existing data. This undefined mode is considered "Version 0", and
- the kernel will now refuse to start these arrays w/o user intervention.
+ However, until an mdadm exists that is able to set a layout in the
+ array, there's no way for the kernel to know which version(s) was used
+ to write the existing data. This undefined mode is considered "Version
+ 0", and the kernel will now refuse to start these arrays w/o user
+ intervention.
  
  The user experience is pretty awful here. A user upgrades to the next
  SRU and all of a sudden their system stops at an (initramfs) prompt. A
  clueful user can spot something like the following in dmesg:
  
  Here's the message which , as you can see from the log in Comment #1, is
  hidden in a ton of other messages:
  
  [ 72.720232] md/raid0:md0: cannot assemble multi-zone RAID0 with 
default_layout setting
  [ 72.728149] md/raid0: please set raid.default_layout to 1 or 2
  [ 72.733979] md: pers->run() failed ...
  mdadm: failed to start array /dev/md0: Unknown error 524
  
  What that is trying to say is that you should determine if your data -
  specifically the data toward the end of your array - was most likely
  written with a pre-3.14 or post-3.14 kernel. Based on that, reboot with
  the kernel parameter raid0.default_layout=1 or raid0.default_layout=2 on
  the kernel command line. And note it should be *raid0.default_layout*
  not *raid.default_layout* as the message says - a fix for that message
  is now queued for stable:
  
  
https://github.com/torvalds/linux/commit/3874d73e06c9b9dc15de0b7382fc223986d75571)
  
  IMHO, we should work with upstream to create a web page that clearly
  walks the user through this process, and update the error message to
  point to that page. I'd also like to see if we can detect this problem
  *before* the user reboots (debconf?) and help the user fix things. e.g.
  "We detected that you have RAID0 arrays that maybe susceptible to a
  corruption problem", guide the user to choosing a layout, and update the
  mdadm initramfs hook to poke the answer in via sysfs before starting the
  array on reboot.
  
  Note that it also seems like we should investigate backporting this to <
  3.14 kernels. Imagine a user switching between the trusty HWE kernel and
  the GA kernel.
  
  References from users of other distros:
  https://blog.icod.de/2019/10/10/caution-kernel-5-3-4-and-raid0-default_layout/
  
https://www.linuxquestions.org/questions/linux-general-1/raid-arrays-not-assembling-4175662774/
  
  [*] Which surprisingly is not the case reported in this bug - the user
  here had a raid0 of 8 identically-sized devices. I suspect there's a bug
  in the detection code somewhere.

-- 
You received this bug notification because you are a member of Ubuntu
Bugs, which is subscribed to Ubuntu.
https://bugs.launchpad.net/bugs/1850540

Title:
  multi-zone raid0 corruption

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/linux/+bug/1850540/+subscriptions

-- 
ubuntu-bugs mailing list
ubuntu-bugs@lists.ubuntu.com
https://lists.ubuntu.com/mailman/listinfo/ubuntu-bugs

[Bug 1850540] Re: multi-zone raid0 corruption

2019-10-30 Thread dann frazier
I've nominated mdadm for this issue as well because it currently does not:
  1) Let you create a new raid0 array w/o first setting a default_layout, which 
will regress behavior in installers, etc.
  2) save the layout version in the raid config, so you need to set 
default_layout - or set individual array layouts via sysfs - on every boot
  3) doesn't allow you to specify a layout when creating a raid0 (--layout is 
permitted for raid5, but not for raid0 yet)

-- 
You received this bug notification because you are a member of Ubuntu
Bugs, which is subscribed to Ubuntu.
https://bugs.launchpad.net/bugs/1850540

Title:
  multi-zone raid0 corruption

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/linux/+bug/1850540/+subscriptions

-- 
ubuntu-bugs mailing list
ubuntu-bugs@lists.ubuntu.com
https://lists.ubuntu.com/mailman/listinfo/ubuntu-bugs