From: "Aneesh Kumar K.V" <aneesh.ku...@linux.vnet.ibm.com> With table_size as second argument, first argument of the function is not the shift value, but rather index into the array. Rename the variable to clarify the same.
Signed-off-by: Aneesh Kumar K.V <aneesh.ku...@linux.vnet.ibm.com> --- arch/powerpc/include/asm/pgtable-ppc64.h | 2 +- arch/powerpc/mm/init_64.c | 16 ++++++++-------- 2 files changed, 9 insertions(+), 9 deletions(-) diff --git a/arch/powerpc/include/asm/pgtable-ppc64.h b/arch/powerpc/include/asm/pgtable-ppc64.h index d51d893..658ba7c 100644 --- a/arch/powerpc/include/asm/pgtable-ppc64.h +++ b/arch/powerpc/include/asm/pgtable-ppc64.h @@ -338,7 +338,7 @@ static inline void __ptep_set_access_flags(pte_t *ptep, pte_t entry) #define pgoff_to_pte(off) ((pte_t) {((off) << PTE_RPN_SHIFT)|_PAGE_FILE}) #define PTE_FILE_MAX_BITS (BITS_PER_LONG - PTE_RPN_SHIFT) -extern void __pgtable_cache_add(unsigned shift, unsigned long table_size, +extern void __pgtable_cache_add(unsigned index, unsigned long table_size, void (*ctor)(void *)); void pgtable_cache_init(void); static inline void pgtable_cache_add(unsigned shift, void (*ctor)(void *)) diff --git a/arch/powerpc/mm/init_64.c b/arch/powerpc/mm/init_64.c index d6df419..2e09ec1 100644 --- a/arch/powerpc/mm/init_64.c +++ b/arch/powerpc/mm/init_64.c @@ -100,7 +100,7 @@ struct kmem_cache *pgtable_cache[MAX_PGTABLE_INDEX_SIZE]; * everything else. Caches created by this function are used for all * the higher level pagetables, and for hugepage pagetables. */ -void __pgtable_cache_add(unsigned int shift, unsigned long table_size, +void __pgtable_cache_add(unsigned int index, unsigned long table_size, void (*ctor)(void *)) { char *name; @@ -110,8 +110,8 @@ void __pgtable_cache_add(unsigned int shift, unsigned long table_size, * the index size in the low bits. Table alignment must be * big enough to fit it. * - * Likewise, hugeapge pagetable pointers contain a (different) - * shift value in the low bits. All tables must be aligned so + * Likewise, hugepage pagetable pointers contain a (different) + * huge page size in the low bits. All tables must be aligned so * as to leave enough 0 bits in the address to contain it. */ unsigned long minalign = max(MAX_PGTABLE_INDEX_SIZE + 1, HUGEPD_SHIFT_MASK + 1); @@ -121,17 +121,17 @@ void __pgtable_cache_add(unsigned int shift, unsigned long table_size, * moment, gcc doesn't seem to recognize is_power_of_2 as a * constant expression, so so much for that. */ BUG_ON(!is_power_of_2(minalign)); - BUG_ON((shift < 1) || (shift > MAX_PGTABLE_INDEX_SIZE)); + BUG_ON((index < 1) || (index > MAX_PGTABLE_INDEX_SIZE)); - if (PGT_CACHE(shift)) + if (PGT_CACHE(index)) return; /* Already have a cache of this size */ align = max_t(unsigned long, align, minalign); - name = kasprintf(GFP_KERNEL, "pgtable-2^%d", shift); + name = kasprintf(GFP_KERNEL, "pgtable-2^%d", index); new = kmem_cache_create(name, table_size, align, 0, ctor); - PGT_CACHE(shift) = new; + PGT_CACHE(index) = new; - pr_debug("Allocated pgtable cache for order %d\n", shift); + pr_debug("Allocated pgtable cache for order %d\n", index); } -- 1.7.10 _______________________________________________ Linuxppc-dev mailing list Linuxppc-dev@lists.ozlabs.org https://lists.ozlabs.org/listinfo/linuxppc-dev