Re: [tip:numa/core] sched/numa/mm: Improve migration
On Thu, Oct 18, 2012 at 10:05:39AM -0700, tip-bot for Peter Zijlstra wrote: > Commit-ID: 713f937655c4b15131b5a0eae4610918a4febe17 > Gitweb: http://git.kernel.org/tip/713f937655c4b15131b5a0eae4610918a4febe17 > Author: Peter Zijlstra > AuthorDate: Fri, 12 Oct 2012 19:30:14 +0200 > Committer: Ingo Molnar > CommitDate: Mon, 15 Oct 2012 14:18:40 +0200 > > sched/numa/mm: Improve migration > > Add THP migration. Extend task_numa_fault() to absorb THP faults. > > [ Would be nice if the gents on Cc: expressed their opinion about > this change. A missing detail might be cgroup page accounting, > plus the fact that some architectures might cache PMD_NONE pmds > in their TLBs, needing some extra TLB magic beyond what we already > do here? ] > I'm travelling for a conference at the moment so will not get the chance to properly review this until I get back. Is there any plan to post the schednuma patches to linux-mm so the full series can be reviewed? I can extract the patches from -tip when I get back but it's still less than ideal from a review standpoint. Superficially, the patch looks ok but as I lack context on what the rest of schednuma looks like I cannot be sure so I'm not going to ack it. Basically this is very similar to __unmap_and_move except it doesn't deal with migration PTEs -- presumably because the PTE is PROT_NONE so it gets queued up behind it. There is a downside of that. With migration PTEs, faults during migration will wait on the PTE. With this approach, I think multiple faults will alloc a hugepage, realise the ptes are no longer the same and back off. It should still work but it's potentially more expensive. Was that considered? Is it deliberate? If so, why? It also feels like the migration part should have been a helper function called unmap_and_move_thp() in migrate.c instead of being buried in mm/huge_memory.c -- Mel Gorman SUSE Labs -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majord...@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/
Re: [tip:numa/core] sched/numa/mm: Improve migration
On Thu, Oct 18, 2012 at 10:05:39AM -0700, tip-bot for Peter Zijlstra wrote: Commit-ID: 713f937655c4b15131b5a0eae4610918a4febe17 Gitweb: http://git.kernel.org/tip/713f937655c4b15131b5a0eae4610918a4febe17 Author: Peter Zijlstra a.p.zijls...@chello.nl AuthorDate: Fri, 12 Oct 2012 19:30:14 +0200 Committer: Ingo Molnar mi...@kernel.org CommitDate: Mon, 15 Oct 2012 14:18:40 +0200 sched/numa/mm: Improve migration Add THP migration. Extend task_numa_fault() to absorb THP faults. [ Would be nice if the gents on Cc: expressed their opinion about this change. A missing detail might be cgroup page accounting, plus the fact that some architectures might cache PMD_NONE pmds in their TLBs, needing some extra TLB magic beyond what we already do here? ] I'm travelling for a conference at the moment so will not get the chance to properly review this until I get back. Is there any plan to post the schednuma patches to linux-mm so the full series can be reviewed? I can extract the patches from -tip when I get back but it's still less than ideal from a review standpoint. Superficially, the patch looks ok but as I lack context on what the rest of schednuma looks like I cannot be sure so I'm not going to ack it. Basically this is very similar to __unmap_and_move except it doesn't deal with migration PTEs -- presumably because the PTE is PROT_NONE so it gets queued up behind it. There is a downside of that. With migration PTEs, faults during migration will wait on the PTE. With this approach, I think multiple faults will alloc a hugepage, realise the ptes are no longer the same and back off. It should still work but it's potentially more expensive. Was that considered? Is it deliberate? If so, why? It also feels like the migration part should have been a helper function called unmap_and_move_thp() in migrate.c instead of being buried in mm/huge_memory.c -- Mel Gorman SUSE Labs -- To unsubscribe from this list: send the line unsubscribe linux-kernel in the body of a message to majord...@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/
Re: [tip:numa/core] sched/numa/mm: Improve migration
* Johannes Weiner wrote: > On Thu, Oct 18, 2012 at 10:05:39AM -0700, tip-bot for Peter Zijlstra wrote: > > Commit-ID: 713f937655c4b15131b5a0eae4610918a4febe17 > > Gitweb: > > http://git.kernel.org/tip/713f937655c4b15131b5a0eae4610918a4febe17 > > Author: Peter Zijlstra > > AuthorDate: Fri, 12 Oct 2012 19:30:14 +0200 > > Committer: Ingo Molnar > > CommitDate: Mon, 15 Oct 2012 14:18:40 +0200 > > > > sched/numa/mm: Improve migration > > > > Add THP migration. Extend task_numa_fault() to absorb THP faults. > > > > [ Would be nice if the gents on Cc: expressed their opinion about > > this change. A missing detail might be cgroup page accounting, > > plus the fact that some architectures might cache PMD_NONE pmds > > in their TLBs, needing some extra TLB magic beyond what we already > > do here? ] > > Looks good to me, the cgroup fixup should be easy enough as well > (added the calls inline below). > > Of course I'm banging my head into a wall for not seeing earlier > through the existing migration path how easy this could be. It would > be great for compaction to have this fastpath in the traditional > migration code too. > > Right now, unlike the traditional migration path, this breaks COW for > every migration, but maybe you don't care about shared pages in the > first place. And fixing that should be nothing more than grabbing the > anon_vma lock and using rmap to switch more than one pmd over, right? > > It won't work for pages in swap, which is only a future problem. > > It's slightly ugly that migrate_page_copy() actually modifies the > existing page (deactivation, munlock) when you end up having to revert > back to it. > > The new page needs to be PageUptodate. > > > + task_numa_placement(); > > + > > + new_page = alloc_pages_node(node, > > + (GFP_TRANSHUGE | GFP_THISNODE) & ~(__GFP_NO_KSWAPD | __GFP_WAIT), > > + HPAGE_PMD_ORDER); > > + > > + WARN_ON(PageLRU(new_page)); This WARN_ON() is somewhat problematic in OOM or OOLNM situations, so I removed it ;-) > > + > > + if (!new_page) > > + goto alloc_fail; > > mem_cgroup_prepare_migration(page, new_page, ); > > > + lru = PageLRU(page); > > + > > + if (lru && isolate_lru_page(page)) /* does an implicit get_page() */ > > + goto alloc_fail; > > + > > + if (!trylock_page(new_page)) > > + BUG(); > > + > > + /* anon mapping, we can simply copy page->mapping to the new page: */ > > + new_page->mapping = page->mapping; > > + new_page->index = page->index; > > + > > + migrate_page_copy(new_page, page); > > + > > + WARN_ON(PageLRU(new_page)); > > > > -do_fixup: > > spin_lock(>page_table_lock); > > - if (unlikely(!pmd_same(*pmd, entry))) > > - goto out_unlock; > > -#endif > > + if (unlikely(!pmd_same(*pmd, entry))) { > > + spin_unlock(>page_table_lock); > > + if (lru) > > + putback_lru_page(page); > > > > - /* change back to regular protection */ > > - entry = pmd_modify(entry, vma->vm_page_prot); > > - if (pmdp_set_access_flags(vma, haddr, pmd, entry, 1)) > > - update_mmu_cache(vma, address, entry); > > + unlock_page(new_page); > > + ClearPageActive(new_page); /* Set by migrate_page_copy() */ > > + new_page->mapping = NULL; > > + put_page(new_page); /* Free it */ > > > > -out_unlock: > > + unlock_page(page); > > + put_page(page); /* Drop the local reference */ > > + > > + return; > > + } > > + > > + entry = mk_pmd(new_page, vma->vm_page_prot); > > + entry = maybe_pmd_mkwrite(pmd_mkdirty(entry), vma); > > + entry = pmd_mkhuge(entry); > > + > > + page_add_new_anon_rmap(new_page, vma, haddr); > > + > > + set_pmd_at(mm, haddr, pmd, entry); > > + update_mmu_cache(vma, address, entry); > > + page_remove_rmap(page); > > spin_unlock(>page_table_lock); > > - if (page) > > + > > + put_page(page); /* Drop the rmap reference */ > > + > > + task_numa_fault(node, HPAGE_PMD_NR); > > + > > + if (lru) > > + put_page(page); /* drop the LRU isolation reference */ > > + > > + unlock_page(new_page); > > mem_cgroup_end_migration(memcg, page, new_page, true); > > > + unlock_page(page); > > + put_page(page); /* Drop the local reference */ > > + > > + return; > > + > > +alloc_fail: > > + if (new_page) > > + put_page(new_page); > mem_cgroup_end_migration(memcg, page, new_page, false); > } > > > + task_numa_fault(page_to_nid(page), HPAGE_PMD_NR); > > + unlock_page(page); > > + > > + spin_lock(>page_table_lock); > > + if (unlikely(!pmd_same(*pmd, entry))) { > > put_page(page); > > + page = NULL; > > + goto unlock; > > + } > > + goto fixup; > > } Cool! Would any of the gents be interested in turning the suggestions above into a
Re: [tip:numa/core] sched/numa/mm: Improve migration
* Johannes Weiner han...@cmpxchg.org wrote: On Thu, Oct 18, 2012 at 10:05:39AM -0700, tip-bot for Peter Zijlstra wrote: Commit-ID: 713f937655c4b15131b5a0eae4610918a4febe17 Gitweb: http://git.kernel.org/tip/713f937655c4b15131b5a0eae4610918a4febe17 Author: Peter Zijlstra a.p.zijls...@chello.nl AuthorDate: Fri, 12 Oct 2012 19:30:14 +0200 Committer: Ingo Molnar mi...@kernel.org CommitDate: Mon, 15 Oct 2012 14:18:40 +0200 sched/numa/mm: Improve migration Add THP migration. Extend task_numa_fault() to absorb THP faults. [ Would be nice if the gents on Cc: expressed their opinion about this change. A missing detail might be cgroup page accounting, plus the fact that some architectures might cache PMD_NONE pmds in their TLBs, needing some extra TLB magic beyond what we already do here? ] Looks good to me, the cgroup fixup should be easy enough as well (added the calls inline below). Of course I'm banging my head into a wall for not seeing earlier through the existing migration path how easy this could be. It would be great for compaction to have this fastpath in the traditional migration code too. Right now, unlike the traditional migration path, this breaks COW for every migration, but maybe you don't care about shared pages in the first place. And fixing that should be nothing more than grabbing the anon_vma lock and using rmap to switch more than one pmd over, right? It won't work for pages in swap, which is only a future problem. It's slightly ugly that migrate_page_copy() actually modifies the existing page (deactivation, munlock) when you end up having to revert back to it. The new page needs to be PageUptodate. + task_numa_placement(); + + new_page = alloc_pages_node(node, + (GFP_TRANSHUGE | GFP_THISNODE) ~(__GFP_NO_KSWAPD | __GFP_WAIT), + HPAGE_PMD_ORDER); + + WARN_ON(PageLRU(new_page)); This WARN_ON() is somewhat problematic in OOM or OOLNM situations, so I removed it ;-) + + if (!new_page) + goto alloc_fail; mem_cgroup_prepare_migration(page, new_page, memcg); + lru = PageLRU(page); + + if (lru isolate_lru_page(page)) /* does an implicit get_page() */ + goto alloc_fail; + + if (!trylock_page(new_page)) + BUG(); + + /* anon mapping, we can simply copy page-mapping to the new page: */ + new_page-mapping = page-mapping; + new_page-index = page-index; + + migrate_page_copy(new_page, page); + + WARN_ON(PageLRU(new_page)); -do_fixup: spin_lock(mm-page_table_lock); - if (unlikely(!pmd_same(*pmd, entry))) - goto out_unlock; -#endif + if (unlikely(!pmd_same(*pmd, entry))) { + spin_unlock(mm-page_table_lock); + if (lru) + putback_lru_page(page); - /* change back to regular protection */ - entry = pmd_modify(entry, vma-vm_page_prot); - if (pmdp_set_access_flags(vma, haddr, pmd, entry, 1)) - update_mmu_cache(vma, address, entry); + unlock_page(new_page); + ClearPageActive(new_page); /* Set by migrate_page_copy() */ + new_page-mapping = NULL; + put_page(new_page); /* Free it */ -out_unlock: + unlock_page(page); + put_page(page); /* Drop the local reference */ + + return; + } + + entry = mk_pmd(new_page, vma-vm_page_prot); + entry = maybe_pmd_mkwrite(pmd_mkdirty(entry), vma); + entry = pmd_mkhuge(entry); + + page_add_new_anon_rmap(new_page, vma, haddr); + + set_pmd_at(mm, haddr, pmd, entry); + update_mmu_cache(vma, address, entry); + page_remove_rmap(page); spin_unlock(mm-page_table_lock); - if (page) + + put_page(page); /* Drop the rmap reference */ + + task_numa_fault(node, HPAGE_PMD_NR); + + if (lru) + put_page(page); /* drop the LRU isolation reference */ + + unlock_page(new_page); mem_cgroup_end_migration(memcg, page, new_page, true); + unlock_page(page); + put_page(page); /* Drop the local reference */ + + return; + +alloc_fail: + if (new_page) + put_page(new_page); mem_cgroup_end_migration(memcg, page, new_page, false); } + task_numa_fault(page_to_nid(page), HPAGE_PMD_NR); + unlock_page(page); + + spin_lock(mm-page_table_lock); + if (unlikely(!pmd_same(*pmd, entry))) { put_page(page); + page = NULL; + goto unlock; + } + goto fixup; } Cool! Would any of the gents be interested in turning the suggestions above into a suitable patch against this tree: git://git.kernel.org/pub/scm/linux/kernel/git/tip/tip.git numa/core ? Thanks a ton! Ingo -- To unsubscribe from this list: send the line unsubscribe
Re: [tip:numa/core] sched/numa/mm: Improve migration
On Fri, 2012-10-19 at 09:51 -0400, Johannes Weiner wrote: > It's slightly ugly that migrate_page_copy() actually modifies the > existing page (deactivation, munlock) when you end up having to revert > back to it. The worst is actually calling copy_huge_page() on a THP.. it seems to work though ;-) -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majord...@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/
Re: [tip:numa/core] sched/numa/mm: Improve migration
On Fri, 2012-10-19 at 09:51 -0400, Johannes Weiner wrote: > Right now, unlike the traditional migration path, this breaks COW for > every migration, but maybe you don't care about shared pages in the > first place. And fixing that should be nothing more than grabbing the > anon_vma lock and using rmap to switch more than one pmd over, right? > This patch was very much about getting _something_ rather than caring (too much) about all the weird corner cases. And yes, your suggestion sounds about right. -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majord...@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/
Re: [tip:numa/core] sched/numa/mm: Improve migration
On Fri, 2012-10-19 at 09:51 -0400, Johannes Weiner wrote: > Of course I'm banging my head into a wall for not seeing earlier > through the existing migration path how easy this could be. There's a reason I keep promoting the idea of 'someone' rewriting all that page-migration code :-) I forever get lost in there. Also note that the proposed code will do 'wasted' work in case the THP page gets split from under us, given that splits are relatively rare (and if they're not, we should make them so) this didn't seem a problem. Also, this code very much relies on our PROT_NONE marking, it avoids the whole migration-PTE dance usually done, further the assumption that THP pages are anonymous only did help keep it simpler -- if someone 'fixes' that this needs more TLC. -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majord...@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/
Re: [tip:numa/core] sched/numa/mm: Improve migration
On Thu, Oct 18, 2012 at 10:05:39AM -0700, tip-bot for Peter Zijlstra wrote: > Commit-ID: 713f937655c4b15131b5a0eae4610918a4febe17 > Gitweb: http://git.kernel.org/tip/713f937655c4b15131b5a0eae4610918a4febe17 > Author: Peter Zijlstra > AuthorDate: Fri, 12 Oct 2012 19:30:14 +0200 > Committer: Ingo Molnar > CommitDate: Mon, 15 Oct 2012 14:18:40 +0200 > > sched/numa/mm: Improve migration > > Add THP migration. Extend task_numa_fault() to absorb THP faults. > > [ Would be nice if the gents on Cc: expressed their opinion about > this change. A missing detail might be cgroup page accounting, > plus the fact that some architectures might cache PMD_NONE pmds > in their TLBs, needing some extra TLB magic beyond what we already > do here? ] Looks good to me, the cgroup fixup should be easy enough as well (added the calls inline below). Of course I'm banging my head into a wall for not seeing earlier through the existing migration path how easy this could be. It would be great for compaction to have this fastpath in the traditional migration code too. Right now, unlike the traditional migration path, this breaks COW for every migration, but maybe you don't care about shared pages in the first place. And fixing that should be nothing more than grabbing the anon_vma lock and using rmap to switch more than one pmd over, right? It won't work for pages in swap, which is only a future problem. It's slightly ugly that migrate_page_copy() actually modifies the existing page (deactivation, munlock) when you end up having to revert back to it. The new page needs to be PageUptodate. > + task_numa_placement(); > + > + new_page = alloc_pages_node(node, > + (GFP_TRANSHUGE | GFP_THISNODE) & ~(__GFP_NO_KSWAPD | __GFP_WAIT), > + HPAGE_PMD_ORDER); > + > + WARN_ON(PageLRU(new_page)); > + > + if (!new_page) > + goto alloc_fail; mem_cgroup_prepare_migration(page, new_page, ); > + lru = PageLRU(page); > + > + if (lru && isolate_lru_page(page)) /* does an implicit get_page() */ > + goto alloc_fail; > + > + if (!trylock_page(new_page)) > + BUG(); > + > + /* anon mapping, we can simply copy page->mapping to the new page: */ > + new_page->mapping = page->mapping; > + new_page->index = page->index; > + > + migrate_page_copy(new_page, page); > + > + WARN_ON(PageLRU(new_page)); > > -do_fixup: > spin_lock(>page_table_lock); > - if (unlikely(!pmd_same(*pmd, entry))) > - goto out_unlock; > -#endif > + if (unlikely(!pmd_same(*pmd, entry))) { > + spin_unlock(>page_table_lock); > + if (lru) > + putback_lru_page(page); > > - /* change back to regular protection */ > - entry = pmd_modify(entry, vma->vm_page_prot); > - if (pmdp_set_access_flags(vma, haddr, pmd, entry, 1)) > - update_mmu_cache(vma, address, entry); > + unlock_page(new_page); > + ClearPageActive(new_page); /* Set by migrate_page_copy() */ > + new_page->mapping = NULL; > + put_page(new_page); /* Free it */ > > -out_unlock: > + unlock_page(page); > + put_page(page); /* Drop the local reference */ > + > + return; > + } > + > + entry = mk_pmd(new_page, vma->vm_page_prot); > + entry = maybe_pmd_mkwrite(pmd_mkdirty(entry), vma); > + entry = pmd_mkhuge(entry); > + > + page_add_new_anon_rmap(new_page, vma, haddr); > + > + set_pmd_at(mm, haddr, pmd, entry); > + update_mmu_cache(vma, address, entry); > + page_remove_rmap(page); > spin_unlock(>page_table_lock); > - if (page) > + > + put_page(page); /* Drop the rmap reference */ > + > + task_numa_fault(node, HPAGE_PMD_NR); > + > + if (lru) > + put_page(page); /* drop the LRU isolation reference */ > + > + unlock_page(new_page); mem_cgroup_end_migration(memcg, page, new_page, true); > + unlock_page(page); > + put_page(page); /* Drop the local reference */ > + > + return; > + > +alloc_fail: > + if (new_page) > + put_page(new_page); mem_cgroup_end_migration(memcg, page, new_page, false); } > + task_numa_fault(page_to_nid(page), HPAGE_PMD_NR); > + unlock_page(page); > + > + spin_lock(>page_table_lock); > + if (unlikely(!pmd_same(*pmd, entry))) { > put_page(page); > + page = NULL; > + goto unlock; > + } > + goto fixup; > } > > int copy_huge_pmd(struct mm_struct *dst_mm, struct mm_struct *src_mm, -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majord...@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/
Re: [tip:numa/core] sched/numa/mm: Improve migration
On Thu, Oct 18, 2012 at 10:05:39AM -0700, tip-bot for Peter Zijlstra wrote: Commit-ID: 713f937655c4b15131b5a0eae4610918a4febe17 Gitweb: http://git.kernel.org/tip/713f937655c4b15131b5a0eae4610918a4febe17 Author: Peter Zijlstra a.p.zijls...@chello.nl AuthorDate: Fri, 12 Oct 2012 19:30:14 +0200 Committer: Ingo Molnar mi...@kernel.org CommitDate: Mon, 15 Oct 2012 14:18:40 +0200 sched/numa/mm: Improve migration Add THP migration. Extend task_numa_fault() to absorb THP faults. [ Would be nice if the gents on Cc: expressed their opinion about this change. A missing detail might be cgroup page accounting, plus the fact that some architectures might cache PMD_NONE pmds in their TLBs, needing some extra TLB magic beyond what we already do here? ] Looks good to me, the cgroup fixup should be easy enough as well (added the calls inline below). Of course I'm banging my head into a wall for not seeing earlier through the existing migration path how easy this could be. It would be great for compaction to have this fastpath in the traditional migration code too. Right now, unlike the traditional migration path, this breaks COW for every migration, but maybe you don't care about shared pages in the first place. And fixing that should be nothing more than grabbing the anon_vma lock and using rmap to switch more than one pmd over, right? It won't work for pages in swap, which is only a future problem. It's slightly ugly that migrate_page_copy() actually modifies the existing page (deactivation, munlock) when you end up having to revert back to it. The new page needs to be PageUptodate. + task_numa_placement(); + + new_page = alloc_pages_node(node, + (GFP_TRANSHUGE | GFP_THISNODE) ~(__GFP_NO_KSWAPD | __GFP_WAIT), + HPAGE_PMD_ORDER); + + WARN_ON(PageLRU(new_page)); + + if (!new_page) + goto alloc_fail; mem_cgroup_prepare_migration(page, new_page, memcg); + lru = PageLRU(page); + + if (lru isolate_lru_page(page)) /* does an implicit get_page() */ + goto alloc_fail; + + if (!trylock_page(new_page)) + BUG(); + + /* anon mapping, we can simply copy page-mapping to the new page: */ + new_page-mapping = page-mapping; + new_page-index = page-index; + + migrate_page_copy(new_page, page); + + WARN_ON(PageLRU(new_page)); -do_fixup: spin_lock(mm-page_table_lock); - if (unlikely(!pmd_same(*pmd, entry))) - goto out_unlock; -#endif + if (unlikely(!pmd_same(*pmd, entry))) { + spin_unlock(mm-page_table_lock); + if (lru) + putback_lru_page(page); - /* change back to regular protection */ - entry = pmd_modify(entry, vma-vm_page_prot); - if (pmdp_set_access_flags(vma, haddr, pmd, entry, 1)) - update_mmu_cache(vma, address, entry); + unlock_page(new_page); + ClearPageActive(new_page); /* Set by migrate_page_copy() */ + new_page-mapping = NULL; + put_page(new_page); /* Free it */ -out_unlock: + unlock_page(page); + put_page(page); /* Drop the local reference */ + + return; + } + + entry = mk_pmd(new_page, vma-vm_page_prot); + entry = maybe_pmd_mkwrite(pmd_mkdirty(entry), vma); + entry = pmd_mkhuge(entry); + + page_add_new_anon_rmap(new_page, vma, haddr); + + set_pmd_at(mm, haddr, pmd, entry); + update_mmu_cache(vma, address, entry); + page_remove_rmap(page); spin_unlock(mm-page_table_lock); - if (page) + + put_page(page); /* Drop the rmap reference */ + + task_numa_fault(node, HPAGE_PMD_NR); + + if (lru) + put_page(page); /* drop the LRU isolation reference */ + + unlock_page(new_page); mem_cgroup_end_migration(memcg, page, new_page, true); + unlock_page(page); + put_page(page); /* Drop the local reference */ + + return; + +alloc_fail: + if (new_page) + put_page(new_page); mem_cgroup_end_migration(memcg, page, new_page, false); } + task_numa_fault(page_to_nid(page), HPAGE_PMD_NR); + unlock_page(page); + + spin_lock(mm-page_table_lock); + if (unlikely(!pmd_same(*pmd, entry))) { put_page(page); + page = NULL; + goto unlock; + } + goto fixup; } int copy_huge_pmd(struct mm_struct *dst_mm, struct mm_struct *src_mm, -- To unsubscribe from this list: send the line unsubscribe linux-kernel in the body of a message to majord...@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/
Re: [tip:numa/core] sched/numa/mm: Improve migration
On Fri, 2012-10-19 at 09:51 -0400, Johannes Weiner wrote: Of course I'm banging my head into a wall for not seeing earlier through the existing migration path how easy this could be. There's a reason I keep promoting the idea of 'someone' rewriting all that page-migration code :-) I forever get lost in there. Also note that the proposed code will do 'wasted' work in case the THP page gets split from under us, given that splits are relatively rare (and if they're not, we should make them so) this didn't seem a problem. Also, this code very much relies on our PROT_NONE marking, it avoids the whole migration-PTE dance usually done, further the assumption that THP pages are anonymous only did help keep it simpler -- if someone 'fixes' that this needs more TLC. -- To unsubscribe from this list: send the line unsubscribe linux-kernel in the body of a message to majord...@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/
Re: [tip:numa/core] sched/numa/mm: Improve migration
On Fri, 2012-10-19 at 09:51 -0400, Johannes Weiner wrote: Right now, unlike the traditional migration path, this breaks COW for every migration, but maybe you don't care about shared pages in the first place. And fixing that should be nothing more than grabbing the anon_vma lock and using rmap to switch more than one pmd over, right? This patch was very much about getting _something_ rather than caring (too much) about all the weird corner cases. And yes, your suggestion sounds about right. -- To unsubscribe from this list: send the line unsubscribe linux-kernel in the body of a message to majord...@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/
Re: [tip:numa/core] sched/numa/mm: Improve migration
On Fri, 2012-10-19 at 09:51 -0400, Johannes Weiner wrote: It's slightly ugly that migrate_page_copy() actually modifies the existing page (deactivation, munlock) when you end up having to revert back to it. The worst is actually calling copy_huge_page() on a THP.. it seems to work though ;-) -- To unsubscribe from this list: send the line unsubscribe linux-kernel in the body of a message to majord...@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/