This combines the non-sleep-capable RCU locking of #v9 with a seqlock so the mmu notifier fast path will require zero cacheline writes/bouncing while still providing mmu_notifier_unregister and allowing to schedule inside the mmu notifier methods. If we drop mmu_notifier_unregister we can as well drop all seqlock and rcu_read_lock()s. But this locking scheme combination is sexy enough and 100% scalable (the mmu_notifier_list cacheline will be preloaded anyway and that will most certainly include the sequence number value in l1 for free even in Christoph's NUMA systems) so IMHO it worth to keep mmu_notifier_unregister.
Signed-off-by: Andrea Arcangeli <[EMAIL PROTECTED]> diff --git a/include/linux/mm_types.h b/include/linux/mm_types.h --- a/include/linux/mm_types.h +++ b/include/linux/mm_types.h @@ -10,6 +10,7 @@ #include <linux/rbtree.h> #include <linux/rwsem.h> #include <linux/completion.h> +#include <linux/seqlock.h> #include <asm/page.h> #include <asm/mmu.h> @@ -230,6 +231,7 @@ struct mm_struct { #endif #ifdef CONFIG_MMU_NOTIFIER struct hlist_head mmu_notifier_list; + seqlock_t mmu_notifier_lock; #endif }; diff --git a/include/linux/mmu_notifier.h b/include/linux/mmu_notifier.h --- a/include/linux/mmu_notifier.h +++ b/include/linux/mmu_notifier.h @@ -130,6 +130,7 @@ static inline void mmu_notifier_mm_init( static inline void mmu_notifier_mm_init(struct mm_struct *mm) { INIT_HLIST_HEAD(&mm->mmu_notifier_list); + seqlock_init(&mm->mmu_notifier_lock); } diff --git a/mm/mmu_notifier.c b/mm/mmu_notifier.c --- a/mm/mmu_notifier.c +++ b/mm/mmu_notifier.c @@ -20,7 +20,9 @@ void __mmu_notifier_release(struct mm_st void __mmu_notifier_release(struct mm_struct *mm) { struct mmu_notifier *mn; + unsigned seq; + seq = read_seqbegin(&mm->mmu_notifier_lock); while (unlikely(!hlist_empty(&mm->mmu_notifier_list))) { mn = hlist_entry(mm->mmu_notifier_list.first, struct mmu_notifier, @@ -28,6 +30,7 @@ void __mmu_notifier_release(struct mm_st hlist_del(&mn->hlist); if (mn->ops->release) mn->ops->release(mn, mm); + BUG_ON(read_seqretry(&mm->mmu_notifier_lock, seq)); } } @@ -42,11 +45,19 @@ int __mmu_notifier_clear_flush_young(str struct mmu_notifier *mn; struct hlist_node *n; int young = 0; + unsigned seq; rcu_read_lock(); +restart: + seq = read_seqbegin(&mm->mmu_notifier_lock); hlist_for_each_entry_rcu(mn, n, &mm->mmu_notifier_list, hlist) { - if (mn->ops->clear_flush_young) + if (mn->ops->clear_flush_young) { + rcu_read_unlock(); young |= mn->ops->clear_flush_young(mn, mm, address); + rcu_read_lock(); + } + if (read_seqretry(&mm->mmu_notifier_lock, seq)) + goto restart; } rcu_read_unlock(); @@ -58,11 +69,19 @@ void __mmu_notifier_invalidate_page(stru { struct mmu_notifier *mn; struct hlist_node *n; + unsigned seq; rcu_read_lock(); +restart: + seq = read_seqbegin(&mm->mmu_notifier_lock); hlist_for_each_entry_rcu(mn, n, &mm->mmu_notifier_list, hlist) { - if (mn->ops->invalidate_page) + if (mn->ops->invalidate_page) { + rcu_read_unlock(); mn->ops->invalidate_page(mn, mm, address); + rcu_read_lock(); + } + if (read_seqretry(&mm->mmu_notifier_lock, seq)) + goto restart; } rcu_read_unlock(); } @@ -72,11 +91,19 @@ void __mmu_notifier_invalidate_range_beg { struct mmu_notifier *mn; struct hlist_node *n; + unsigned seq; rcu_read_lock(); +restart: + seq = read_seqbegin(&mm->mmu_notifier_lock); hlist_for_each_entry_rcu(mn, n, &mm->mmu_notifier_list, hlist) { - if (mn->ops->invalidate_range_begin) + if (mn->ops->invalidate_range_begin) { + rcu_read_unlock(); mn->ops->invalidate_range_begin(mn, mm, start, end); + rcu_read_lock(); + } + if (read_seqretry(&mm->mmu_notifier_lock, seq)) + goto restart; } rcu_read_unlock(); } @@ -86,11 +113,19 @@ void __mmu_notifier_invalidate_range_end { struct mmu_notifier *mn; struct hlist_node *n; + unsigned seq; rcu_read_lock(); +restart: + seq = read_seqbegin(&mm->mmu_notifier_lock); hlist_for_each_entry_rcu(mn, n, &mm->mmu_notifier_list, hlist) { - if (mn->ops->invalidate_range_end) + if (mn->ops->invalidate_range_end) { + rcu_read_unlock(); mn->ops->invalidate_range_end(mn, mm, start, end); + rcu_read_lock(); + } + if (read_seqretry(&mm->mmu_notifier_lock, seq)) + goto restart; } rcu_read_unlock(); } @@ -103,12 +138,20 @@ void __mmu_notifier_invalidate_range_end */ void mmu_notifier_register(struct mmu_notifier *mn, struct mm_struct *mm) { + /* no need of seqlock for hlist_add_head_rcu */ hlist_add_head_rcu(&mn->hlist, &mm->mmu_notifier_list); } EXPORT_SYMBOL_GPL(mmu_notifier_register); void mmu_notifier_unregister(struct mmu_notifier *mn, struct mm_struct *mm) { + /* + * The seqlock tracks if a hlist_del_rcu happens while a + * notifier method is scheduling and in such a case the "mn" + * memory may have been freed by the time the method returns. + */ + write_seqlock(&mm->mmu_notifier_lock); hlist_del_rcu(&mn->hlist); + write_sequnlock(&mm->mmu_notifier_lock); } EXPORT_SYMBOL_GPL(mmu_notifier_unregister); ------------------------------------------------------------------------- This SF.net email is sponsored by: Microsoft Defy all challenges. Microsoft(R) Visual Studio 2008. http://clk.atdmt.com/MRT/go/vse0120000070mrt/direct/01/ _______________________________________________ kvm-devel mailing list kvm-devel@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/kvm-devel