On Tue, 8 Jul 2025, Nilay Shroff wrote:

> 
> 
> On 7/7/25 6:41 PM, John Garry wrote:
> > The atomic write unit max value is limited by any stacked device stripe
> > size.
> > 
> > It is required that the atomic write unit is a power-of-2 factor of the
> > stripe size.
> > 
> > Currently we use io_min limit to hold the stripe size, and check for a
> > io_min <= SECTOR_SIZE when deciding if we have a striped stacked device.
> > 
> > Nilay reports that this causes a problem when the physical block size is
> > greater than SECTOR_SIZE [0].
> > 
> > Furthermore, io_min may be mutated when stacking devices, and this makes
> > it a poor candidate to hold the stripe size. Such an example (of when
> > io_min may change) would be when the io_min is less than the physical
> > block size.
> > 
> > Use chunk_sectors to hold the stripe size, which is more appropriate.
> > 
> > [0] 
> > https://lore.kernel.org/linux-block/888f3b1d-7817-4007-b3b3-1a2ea04df...@linux.ibm.com/T/#mecca17129f72811137d3c2f1e477634e77f06781
> > 
> > Signed-off-by: John Garry <john.g.ga...@oracle.com>
> > ---
> >  block/blk-settings.c | 58 ++++++++++++++++++++++++++------------------
> >  1 file changed, 35 insertions(+), 23 deletions(-)
> > 
> > diff --git a/block/blk-settings.c b/block/blk-settings.c
> > index 761c6ccf5af7..3259cfac5d0d 100644
> > --- a/block/blk-settings.c
> > +++ b/block/blk-settings.c
> > @@ -597,41 +597,52 @@ static bool 
> > blk_stack_atomic_writes_boundary_head(struct queue_limits *t,
> >     return true;
> >  }
> >  
> > -
> > -/* Check stacking of first bottom device */
> > -static bool blk_stack_atomic_writes_head(struct queue_limits *t,
> > -                           struct queue_limits *b)
> > +static void blk_stack_atomic_writes_chunk_sectors(struct queue_limits *t)
> >  {
> > -   if (b->atomic_write_hw_boundary &&
> > -       !blk_stack_atomic_writes_boundary_head(t, b))
> > -           return false;
> > +   unsigned int chunk_sectors = t->chunk_sectors, chunk_bytes;
> >  
> > -   if (t->io_min <= SECTOR_SIZE) {
> > -           /* No chunk sectors, so use bottom device values directly */
> > -           t->atomic_write_hw_unit_max = b->atomic_write_hw_unit_max;
> > -           t->atomic_write_hw_unit_min = b->atomic_write_hw_unit_min;
> > -           t->atomic_write_hw_max = b->atomic_write_hw_max;
> > -           return true;
> > -   }
> > +   if (!chunk_sectors)
> > +           return;
> > +
> > +   /*
> > +    * If chunk sectors is so large that its value in bytes overflows
> > +    * UINT_MAX, then just shift it down so it definitely will fit.
> > +    * We don't support atomic writes of such a large size anyway.
> > +    */
> > +   if ((unsigned long)chunk_sectors << SECTOR_SHIFT > UINT_MAX)
> > +           chunk_bytes = chunk_sectors;
> > +   else
> > +           chunk_bytes = chunk_sectors << SECTOR_SHIFT;

Why do we cast it to unsigned long? unsigned long is 32-bit on 32-bit 
machines, so the code will not detect the overflow in that case. We should 
cast it to unsigned long long (or uint64_t).

Mikulas


Reply via email to