From: Thomas Gleixner <[email protected]> CONFIG_PREEMPT_COUNT is now unconditionally enabled and will be removed. Cleanup the leftovers before doing so.
Signed-off-by: Thomas Gleixner <[email protected]> Cc: Andrew Morton <[email protected]> Cc: [email protected] Signed-off-by: Uladzislau Rezki (Sony) <[email protected]> --- include/linux/pagemap.h | 4 +--- 1 file changed, 1 insertion(+), 3 deletions(-) diff --git a/include/linux/pagemap.h b/include/linux/pagemap.h index c77b7c31b2e4..cbfbe2bcca75 100644 --- a/include/linux/pagemap.h +++ b/include/linux/pagemap.h @@ -204,9 +204,7 @@ void release_pages(struct page **pages, int nr); static inline int __page_cache_add_speculative(struct page *page, int count) { #ifdef CONFIG_TINY_RCU -# ifdef CONFIG_PREEMPT_COUNT - VM_BUG_ON(!in_atomic() && !irqs_disabled()); -# endif + VM_BUG_ON(preemptible()) /* * Preempt must be disabled here - we rely on rcu_read_lock doing * this for us. -- 2.20.1

