On Fri, Sep 02 2016 at  6:42pm -0400,
Bart Van Assche <[email protected]> wrote:

> However, I think
> in dm_stop_queue() all we need is to wait until queue_rq() has
> finished. How about adding new functions in the block layer core to
> realize this, e.g. something like in the attached (untested) patch?
> Busy looping should be avoided - see also the tests of the new
> "quiescing" flag.
> 
> Thanks,
> 
> Bart.

Comments inlined below.

> From e55a161ee4df7804767ed8faf9ddb698e8852b06 Mon Sep 17 00:00:00 2001
> From: Bart Van Assche <[email protected]>
> Date: Fri, 2 Sep 2016 09:32:17 -0700
> Subject: [PATCH] blk-mq: Introduce blk_mq_quiesce_queue()
> 
> ---
>  block/blk-mq.c         | 50 
> +++++++++++++++++++++++++++++++++++++++++++++++++-
>  include/linux/blk-mq.h |  2 ++
>  include/linux/blkdev.h |  3 +++
>  3 files changed, 54 insertions(+), 1 deletion(-)
> 
> diff --git a/block/blk-mq.c b/block/blk-mq.c
> index 123d1ad..0320cd9 100644
> --- a/block/blk-mq.c
> +++ b/block/blk-mq.c
> @@ -135,6 +135,46 @@ void blk_mq_unfreeze_queue(struct request_queue *q)
>  }
>  EXPORT_SYMBOL_GPL(blk_mq_unfreeze_queue);
>  
> +/**
> + * blk_mq_quiesce_queue - wait until all pending queue_rq calls have finished
> + *
> + * Prevent that new I/O requests are queued and wait until all pending
> + * queue_rq() calls have finished.
> + */
> +void blk_mq_quiesce_queue(struct request_queue *q)
> +{
> +     spin_lock_irq(q->queue_lock);
> +     WARN_ON_ONCE(blk_queue_quiescing(q));
> +     queue_flag_set(QUEUE_FLAG_QUIESCING, q);
> +     spin_unlock_irq(q->queue_lock);
> +
> +     atomic_inc_return(&q->mq_freeze_depth);
> +     blk_mq_run_hw_queues(q, false);
> +     synchronize_rcu();

Why the synchronize_rcu()?

Also, you're effectively open-coding blk_mq_freeze_queue_start() minus
the q->q_usage_counter mgmt.  Why not add a flag to conditionally manage
q->q_usage_counter to blk_mq_freeze_queue_start()?

> +     spin_lock_irq(q->queue_lock);
> +     WARN_ON_ONCE(!blk_queue_quiescing(q));
> +     queue_flag_clear(QUEUE_FLAG_QUIESCING, q);
> +     spin_unlock_irq(q->queue_lock);
> +}
> +EXPORT_SYMBOL_GPL(blk_mq_quiesce_queue);
> +
> +/**
> + * blk_mq_resume_queue - resume request processing
> + */
> +void blk_mq_resume_queue(struct request_queue *q)
> +{
> +     int freeze_depth;
> +
> +     freeze_depth = atomic_dec_return(&q->mq_freeze_depth);
> +     WARN_ON_ONCE(freeze_depth < 0);
> +     if (freeze_depth == 0)
> +             wake_up_all(&q->mq_freeze_wq);

Likewise, here you've open coded blk_mq_unfreeze_queue().  Adding a flag
to conditionally reinit q->q_usage_counter would be better.

But I'm concerned about blk_mq_{quiesce,resume}_queue vs
blk_mq_{freeze,unfreeze}_queue -- e.g. if "freeze" is nested after
"queue" (but before "resume") it would still need the q->q_usage_counter
management.  Your patch as-is would break the blk-mq freeze interface.

> +     blk_mq_run_hw_queues(q, false);
> +}
> +EXPORT_SYMBOL_GPL(blk_mq_resume_queue);
> +
>  void blk_mq_wake_waiters(struct request_queue *q)
>  {
>       struct blk_mq_hw_ctx *hctx;
> @@ -506,6 +546,9 @@ static void blk_mq_requeue_work(struct work_struct *work)
>       struct request *rq, *next;
>       unsigned long flags;
>  
> +     if (blk_queue_quiescing(q))
> +             return;
> +
>       spin_lock_irqsave(&q->requeue_lock, flags);
>       list_splice_init(&q->requeue_list, &rq_list);
>       spin_unlock_irqrestore(&q->requeue_lock, flags);
> @@ -806,6 +849,8 @@ static void __blk_mq_run_hw_queue(struct blk_mq_hw_ctx 
> *hctx)
>        */
>       flush_busy_ctxs(hctx, &rq_list);
>  
> +     rcu_read_lock();
> +
>       /*
>        * If we have previous entries on our dispatch list, grab them
>        * and stuff them at the front for more fair dispatch.
> @@ -888,8 +933,11 @@ static void __blk_mq_run_hw_queue(struct blk_mq_hw_ctx 
> *hctx)
>                *
>                * blk_mq_run_hw_queue() already checks the STOPPED bit
>                **/
> -             blk_mq_run_hw_queue(hctx, true);
> +             if (!blk_queue_quiescing(q))
> +                     blk_mq_run_hw_queue(hctx, true);
>       }
> +
> +     rcu_read_unlock();
>  }
>  
>  /*

Please explain this extra rcu_read_{lock,unlock}

--
dm-devel mailing list
[email protected]
https://www.redhat.com/mailman/listinfo/dm-devel

Reply via email to