For the purposes of this prototype, copy the body of pagevec_lru_move_fn
into __pagevec_lru_add so that it can be modified to use the batch
locking API while leaving all other callers of pagevec_lru_move_fn
unaffected.

Signed-off-by: Daniel Jordan <daniel.m.jor...@oracle.com>
---
 mm/swap.c | 24 +++++++++++++++++++++++-
 1 file changed, 23 insertions(+), 1 deletion(-)

diff --git a/mm/swap.c b/mm/swap.c
index cf6a59f2cad6..2bb28fcb7cc0 100644
--- a/mm/swap.c
+++ b/mm/swap.c
@@ -902,7 +902,29 @@ static void __pagevec_lru_add_fn(struct page *page, struct 
lruvec *lruvec,
  */
 void __pagevec_lru_add(struct pagevec *pvec)
 {
-       pagevec_lru_move_fn(pvec, __pagevec_lru_add_fn, NULL);
+       int i;
+       struct pglist_data *pgdat = NULL;
+       struct lruvec *lruvec;
+       unsigned long flags = 0;
+
+       for (i = 0; i < pagevec_count(pvec); i++) {
+               struct page *page = pvec->pages[i];
+               struct pglist_data *pagepgdat = page_pgdat(page);
+
+               if (pagepgdat != pgdat) {
+                       if (pgdat)
+                               spin_unlock_irqrestore(&pgdat->lru_lock, flags);
+                       pgdat = pagepgdat;
+                       spin_lock_irqsave(&pgdat->lru_lock, flags);
+               }
+
+               lruvec = mem_cgroup_page_lruvec(page, pgdat);
+               __pagevec_lru_add_fn(page, lruvec, NULL);
+       }
+       if (pgdat)
+               spin_unlock_irqrestore(&pgdat->lru_lock, flags);
+       release_pages(pvec->pages, pvec->nr);
+       pagevec_reinit(pvec);
 }
 EXPORT_SYMBOL(__pagevec_lru_add);
 
-- 
2.16.1

Reply via email to