On Wed, 2013-05-29 at 22:47 -0700, Hugh Dickins wrote: > Running my favourite swapping load (repeated make -j20 kernel builds > in tmpfs in parallel with repeated make -j20 kernel builds in ext4 on > loop on tmpfs file, all limited by mem=700M and swap 1.5G) on 3.10-rc > on PowerMac G5, the test dies with corrupted usermem after a few hours. > > Variously, segmentation fault or Binutils assertion fail or gcc Internal > error in either or both builds: usually signs of swapping or TLB flushing > gone wrong. Sometimes the tmpfs build breaks first, sometimes the ext4 on > loop on tmpfs, so at least it looks unrelated to loop. No problem on x86. > > This is 64-bit kernel but 4k pages and old SuSE 11.1 32-bit userspace. > > I've just finished a manual bisection on arch/powerpc/mm (which might > have been a wrong guess, but has paid off): the first bad commit is > 7e74c3921ad9610c0b49f28b8fc69f7480505841 > "powerpc: Fix hpte_decode to use the correct decoding for page sizes".
Ok, I have other reasons to think is wrong. I debugged a case last week where after kexec we still had stale TLB entries, due to the TLB cleanup not working. Thanks for doing that bisection ! I'll investigate ASAP (though it will probably have to wait for tomorrow unless Paul beats me to it) > I don't know if it's actually swapping to swap that's triggering the > problem, or a more general page reclaim or TLB flush problem. I hit > it originally when trying to test Mel Gorman's pagevec series on top > of 3.10-rc; and though I then reproduced it without that series, it > did seem to take much longer: so I have been applying Mel's series to > speed up each step of the bisection. But if I went back again, might > find it was just chance that I hit it sooner with Mel's series than > without. So, you're probably safe to ignore that detail, but I > mention it just in case it turns out to have some relevance. > > Something else peculiar that I've been doing in these runs, may or may > not be relevant: I've been running swapon and swapoff repeatedly in the > background, so that we're doing swapoff even while busy building. > > I probably can't go into much more detail on the test (it's hard > to get the balance right, to be swapping rather than OOMing or just > running without reclaim), but can test any patches you'd like me to > try (though it may take 24 hours for me to report back usefully). I think it's just failing to invalidate the TLB properly. At least one bug I can spot just looking at it: static void native_hpte_invalidate(unsigned long slot, unsigned long vpn, int psize, int ssize, int local) .../... native_lock_hpte(hptep); hpte_v = hptep->v; actual_psize = hpte_actual_psize(hptep, psize); if (actual_psize < 0) { native_unlock_hpte(hptep); local_irq_restore(flags); return; } That's wrong. We must still perform the TLB invalidation even if the hash PTE is empty. In fact, Aneesh, this is a problem with MPSS for your THP work, I just thought about it. The reason is that if a hash bucket gets full, we "evict" a more/less random entry from it. When we do that we don't invalidate the TLB (hpte_remove) because we assume the old translation is still technically "valid". However that means that an hpte_invalidate *must* invalidate the TLB later on even if it's not hitting the right entry in the hash. However, I can see why that cannot work with THP/MPSS since you have no way to know the page size from the PTE anymore.... So my question is, apart from hpte_decode used by kexec, which I will fix by just blowing the whole TLB when not running phyp, why do you need the "actual" size in invalidate and updatepp ? You really can't rely on the size passed by the upper layers ? Cheers, Ben. _______________________________________________ Linuxppc-dev mailing list Linuxppc-dev@lists.ozlabs.org https://lists.ozlabs.org/listinfo/linuxppc-dev