On Wed, 2011-05-18 at 16:05 -0500, Scott Wood wrote: > Loads with non-linear access patterns were producing a very high > ratio of recursive pt faults to regular tlb misses. Rather than > choose between a 4-level table walk or a 1-level virtual page table > lookup, use a hybrid scheme with a virtual linear pmd, followed by a > 2-level lookup in the normal handler. > > This adds about 5 cycles (assuming no cache misses, and e5500 timing) > to a normal TLB miss, but greatly reduces the recursive fault rate > for loads which don't have locality within 2 MiB regions but do have > significant locality within 1 GiB regions. Improvements of close to 50% > were seen on such benchmarks.
Can you publish benchmarks that compare these two with no virtual at all (4 full loads) ? Cheers, Ben. > Signed-off-by: Scott Wood <scottw...@freescale.com> > --- > arch/powerpc/mm/tlb_low_64e.S | 23 +++++++++++++++-------- > 1 files changed, 15 insertions(+), 8 deletions(-) > > diff --git a/arch/powerpc/mm/tlb_low_64e.S b/arch/powerpc/mm/tlb_low_64e.S > index af08922..17726d3 100644 > --- a/arch/powerpc/mm/tlb_low_64e.S > +++ b/arch/powerpc/mm/tlb_low_64e.S > @@ -24,7 +24,7 @@ > #ifdef CONFIG_PPC_64K_PAGES > #define VPTE_PMD_SHIFT (PTE_INDEX_SIZE+1) > #else > -#define VPTE_PMD_SHIFT (PTE_INDEX_SIZE) > +#define VPTE_PMD_SHIFT 0 > #endif > #define VPTE_PUD_SHIFT (VPTE_PMD_SHIFT + PMD_INDEX_SIZE) > #define VPTE_PGD_SHIFT (VPTE_PUD_SHIFT + PUD_INDEX_SIZE) > @@ -185,7 +185,7 @@ normal_tlb_miss: > /* Insert the bottom bits in */ > rlwimi r14,r15,0,16,31 > #else > - rldicl r14,r16,64-(PAGE_SHIFT-3),PAGE_SHIFT-3+4 > + rldicl r14,r16,64-(PMD_SHIFT-3),PMD_SHIFT-3+4 > #endif > sldi r15,r10,60 > clrrdi r14,r14,3 > @@ -202,6 +202,16 @@ MMU_FTR_SECTION_ELSE > ld r14,0(r10) > ALT_MMU_FTR_SECTION_END_IFSET(MMU_FTR_USE_TLBRSRV) > > +#ifndef CONFIG_PPC_64K_PAGES > + rldicl r15,r16,64-PAGE_SHIFT+3,64-PTE_INDEX_SIZE-3 > + clrrdi r15,r15,3 > + > + cmpldi cr0,r14,0 > + beq normal_tlb_miss_access_fault > + > + ldx r14,r14,r15 > +#endif > + > finish_normal_tlb_miss: > /* Check if required permissions are met */ > andc. r15,r11,r14 > @@ -353,14 +363,11 @@ END_MMU_FTR_SECTION_IFSET(MMU_FTR_USE_TLBRSRV) > #ifndef CONFIG_PPC_64K_PAGES > /* Get to PUD entry */ > rldicl r11,r16,64-VPTE_PUD_SHIFT,64-PUD_INDEX_SIZE-3 > - clrrdi r10,r11,3 > - ldx r15,r10,r15 > - cmpldi cr0,r15,0 > - beq virt_page_table_tlb_miss_fault > -#endif /* CONFIG_PPC_64K_PAGES */ > - > +#else > /* Get to PMD entry */ > rldicl r11,r16,64-VPTE_PMD_SHIFT,64-PMD_INDEX_SIZE-3 > +#endif > + > clrrdi r10,r11,3 > ldx r15,r10,r15 > cmpldi cr0,r15,0 _______________________________________________ Linuxppc-dev mailing list Linuxppc-dev@lists.ozlabs.org https://lists.ozlabs.org/listinfo/linuxppc-dev