On Tue, Feb 24, 2015 at 11:51 AM, David Sterba <dste...@suse.cz> wrote:
The waitqueue might miss a wakeup due to memory ordering issues, the
explicit barrier is required unless there's an implicit one.

Thanks for going through these Dave, a few comments below:


Signed-off-by: David Sterba <dste...@suse.cz>
---
 fs/btrfs/dev-replace.c |  9 ++++++++-
 fs/btrfs/raid56.c      | 17 ++++++++++++-----
 fs/btrfs/transaction.c |  5 ++++-
 fs/btrfs/tree-log.c    |  8 ++++++++
 4 files changed, 32 insertions(+), 7 deletions(-)

diff --git a/fs/btrfs/dev-replace.c b/fs/btrfs/dev-replace.c
index 5ec03d999c37..30b8668396e0 100644
--- a/fs/btrfs/dev-replace.c
+++ b/fs/btrfs/dev-replace.c
@@ -451,6 +451,10 @@ static void btrfs_rm_dev_replace_blocked(struct btrfs_fs_info *fs_info) static void btrfs_rm_dev_replace_unblocked(struct btrfs_fs_info *fs_info)
 {
        clear_bit(BTRFS_FS_STATE_DEV_REPLACING, &fs_info->fs_state);
+       /*
+        * Make sure counter is updated before we wake up waiters.
+        */
+       smp_mb();
        if (waitqueue_active(&fs_info->replace_wait))
                wake_up(&fs_info->replace_wait);

This one isn't performance critical, lets just use wake_up directly.


 }
@@ -916,7 +920,10 @@ void btrfs_bio_counter_inc_noblocked(struct btrfs_fs_info *fs_info)
 void btrfs_bio_counter_sub(struct btrfs_fs_info *fs_info, s64 amount)
 {
        percpu_counter_sub(&fs_info->bio_counter, amount);
-
+       /*
+        * Make sure counter is updated before we wake up waiters.
+        */
+       smp_mb();
        if (waitqueue_active(&fs_info->replace_wait))
                wake_up(&fs_info->replace_wait);
 }

This is performance critical. Can we do it only when a replace is actually running?


diff --git a/fs/btrfs/raid56.c b/fs/btrfs/raid56.c
index 5264858ed768..b460e193f324 100644
--- a/fs/btrfs/raid56.c
+++ b/fs/btrfs/raid56.c
@@ -809,11 +809,18 @@ static noinline void unlock_stripe(struct btrfs_raid_bio *rbio)
                        }

                        goto done_nolock;
-               } else  if (waitqueue_active(&h->wait)) {
-                       spin_unlock(&rbio->bio_list_lock);
-                       spin_unlock_irqrestore(&h->lock, flags);
-                       wake_up(&h->wait);
-                       goto done_nolock;
+               } else {
+                       /*
+                        * Make sure counter is updated before we wake up
+                        * waiters.
+                        */
+                       smp_mb();
+                       if (waitqueue_active(&h->wait)) {
+                               spin_unlock(&rbio->bio_list_lock);
+                               spin_unlock_irqrestore(&h->lock, flags);
+                               wake_up(&h->wait);
+                               goto done_nolock;
+                       }
                }
        }
 done:

This one is already protected by the h->lock, we can't miss wakeups.


diff --git a/fs/btrfs/transaction.c b/fs/btrfs/transaction.c
index 038fcf6051e0..90ba0c3c3d0d 100644
--- a/fs/btrfs/transaction.c
+++ b/fs/btrfs/transaction.c
@@ -90,8 +90,11 @@ static void clear_btree_io_tree(struct extent_io_tree *tree)
                /*
                 * btree io trees aren't supposed to have tasks waiting for
                 * changes in the flags of extent states ever.
+                *
+                * Barrier required to make sure counter is updated before we
+                * wake up waiters.
                 */
-               ASSERT(!waitqueue_active(&state->wq));
+               ASSERT(({ smp_mb(); !waitqueue_active(&state->wq); }));
                free_extent_state(state);
                if (need_resched()) {
                        spin_unlock(&tree->lock);

Maybe just one before the while loop? This shouldn't be changing once clear_btree_io_tree is called


diff --git a/fs/btrfs/tree-log.c b/fs/btrfs/tree-log.c
index f96996a1b70c..121df0fe5128 100644
--- a/fs/btrfs/tree-log.c
+++ b/fs/btrfs/tree-log.c
@@ -2739,6 +2739,10 @@ out_wake_log_root:
        atomic_set(&log_root_tree->log_commit[index2], 0);
        mutex_unlock(&log_root_tree->log_mutex);

+       /*
+        * Make sure counter is updated before we wake up waiters.
+        */
+       smp_mb();
        if (waitqueue_active(&log_root_tree->log_commit_wait[index2]))
                wake_up(&log_root_tree->log_commit_wait[index2]);
 out:

this one is protected by the log tree mutex


@@ -2750,6 +2754,10 @@ out:
        atomic_set(&root->log_commit[index1], 0);
        mutex_unlock(&root->log_mutex);

+       /*
+        * Make sure counter is updated before we wake up waiters.
+        */
+       smp_mb();
        if (waitqueue_active(&root->log_commit_wait[index1]))
                wake_up(&root->log_commit_wait[index1]);
        return ret;

This one we still need, but you'll get the barrier for free from mutex_unlock

-chris



--
To unsubscribe from this list: send the line "unsubscribe linux-btrfs" in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html

Reply via email to