On Tue, Mar 13, 2018 at 10:05:55PM +0100, John Ogness wrote:

> > +   rcu_read_lock();                /* to protect parent */
> > +   spin_unlock(&dentry->d_lock);
> > +   parent = READ_ONCE(dentry->d_parent);
> 
> The preceeding line should be removed. We already have a "parent" from
> before we did the most recent trylock().

Nope.  We have parent, yes, but it had been fetched outside of rcu_read_lock().
So the object it used to point to might have been already freed and we
can't do this:

> > +   spin_lock(&parent->d_lock);

To get rid of that reread we'd need to do this:
        rcu_read_lock();
        parent = dentry->d_parent;
        if (IS_ROOT(dentry) || likely(spin_trylock(&parent->d_lock))) {
                rcu_read_unlock();
                return true;
        }
        spin_unlock(&dentry->d_lock);
        spin_lock(&parent->d_lock);
        if (unlikely(parent != dentry->d_parent)) {
                ....

Come to think of that, it might make sense to lift rcu_read_lock() all the
way out of that sucker.  Objections?  Below is the incremental I'd fold into
that commit:

diff --git a/fs/dcache.c b/fs/dcache.c
index f0e73c93182b..0d1dac750c0a 100644
--- a/fs/dcache.c
+++ b/fs/dcache.c
@@ -1000,7 +1000,6 @@ static bool shrink_lock_dentry(struct dentry *dentry)
 
        inode = dentry->d_inode;
        if (inode && unlikely(!spin_trylock(&inode->i_lock))) {
-               rcu_read_lock();        /* to protect inode */
                spin_unlock(&dentry->d_lock);
                spin_lock(&inode->i_lock);
                spin_lock(&dentry->d_lock);
@@ -1009,16 +1008,14 @@ static bool shrink_lock_dentry(struct dentry *dentry)
                /* changed inode means that somebody had grabbed it */
                if (unlikely(inode != dentry->d_inode))
                        goto out;
-               rcu_read_unlock();
        }
 
        parent = dentry->d_parent;
+       /* parent will stay allocated until we drop rcu_read_lock */
        if (IS_ROOT(dentry) || likely(spin_trylock(&parent->d_lock)))
                return true;
 
-       rcu_read_lock();                /* to protect parent */
        spin_unlock(&dentry->d_lock);
-       parent = READ_ONCE(dentry->d_parent);
        spin_lock(&parent->d_lock);
        if (unlikely(parent != dentry->d_parent)) {
                spin_unlock(&parent->d_lock);
@@ -1026,14 +1023,11 @@ static bool shrink_lock_dentry(struct dentry *dentry)
                goto out;
        }
        spin_lock_nested(&dentry->d_lock, DENTRY_D_LOCK_NESTED);
-       if (likely(!dentry->d_lockref.count)) {
-               rcu_read_unlock();
+       if (likely(!dentry->d_lockref.count))
                return true;
-       }
        spin_unlock(&parent->d_lock);
 out:
        spin_unlock(&inode->i_lock);
-       rcu_read_unlock();
        return false;
 }
 
@@ -1044,8 +1038,10 @@ static void shrink_dentry_list(struct list_head *list)
 
                dentry = list_entry(list->prev, struct dentry, d_lru);
                spin_lock(&dentry->d_lock);
+               rcu_read_lock();
                if (!shrink_lock_dentry(dentry)) {
                        bool can_free = false;
+                       rcu_read_unlock();
                        d_shrink_del(dentry);
                        if (dentry->d_lockref.count < 0)
                                can_free = dentry->d_flags & DCACHE_MAY_FREE;
@@ -1054,6 +1050,7 @@ static void shrink_dentry_list(struct list_head *list)
                                dentry_free(dentry);
                        continue;
                }
+               rcu_read_unlock();
                d_shrink_del(dentry);
                parent = dentry->d_parent;
                __dentry_kill(dentry);

Reply via email to