This is a note to let you know that I've just added the patch titled
asm-generic/tlb: Track freeing of page-table directories in struct mmu_gather to the 4.19-stable tree which can be found at: http://www.kernel.org/git/?p=linux/kernel/git/stable/stable-queue.git;a=summary The filename of the patch is: asm-generic-tlb-track-freeing-of-page-table-directories-in-struct-mmu_gather.patch and it can be found in the queue-4.19 subdirectory. If you, or anyone else, feels it should not be added to the stable tree, please let <sta...@vger.kernel.org> know about it. >From foo@baz Mon Jan 4 01:45:29 PM CET 2021 From: Santosh Sivaraj <sant...@fossix.org> Date: Thu, 12 Mar 2020 18:57:35 +0530 Subject: asm-generic/tlb: Track freeing of page-table directories in struct mmu_gather To: <sta...@vger.kernel.org>, linuxppc-dev <linuxppc-dev@lists.ozlabs.org> Cc: Michael Ellerman <m...@ellerman.id.au>, Greg KH <g...@kroah.com>, Sasha Levin <sas...@kernel.org>, Peter Zijlstra <pet...@infradead.org>, Will Deacon <will.dea...@arm.com> Message-ID: <20200312132740.225241-2-sant...@fossix.org> From: Peter Zijlstra <pet...@infradead.org> commit 22a61c3c4f1379ef8b0ce0d5cb78baf3178950e2 upstream Some architectures require different TLB invalidation instructions depending on whether it is only the last-level of page table being changed, or whether there are also changes to the intermediate (directory) entries higher up the tree. Add a new bit to the flags bitfield in struct mmu_gather so that the architecture code can operate accordingly if it's the intermediate levels being invalidated. Signed-off-by: Peter Zijlstra <pet...@infradead.org> Signed-off-by: Will Deacon <will.dea...@arm.com> Cc: <sta...@vger.kernel.org> # 4.19 Signed-off-by: Santosh Sivaraj <sant...@fossix.org> [santosh: prerequisite for tlbflush backports] Signed-off-by: Greg Kroah-Hartman <gre...@linuxfoundation.org> --- include/asm-generic/tlb.h | 31 +++++++++++++++++++++++-------- 1 file changed, 23 insertions(+), 8 deletions(-) --- a/include/asm-generic/tlb.h +++ b/include/asm-generic/tlb.h @@ -97,12 +97,22 @@ struct mmu_gather { #endif unsigned long start; unsigned long end; - /* we are in the middle of an operation to clear - * a full mm and can make some optimizations */ - unsigned int fullmm : 1, - /* we have performed an operation which - * requires a complete flush of the tlb */ - need_flush_all : 1; + /* + * we are in the middle of an operation to clear + * a full mm and can make some optimizations + */ + unsigned int fullmm : 1; + + /* + * we have performed an operation which + * requires a complete flush of the tlb + */ + unsigned int need_flush_all : 1; + + /* + * we have removed page directories + */ + unsigned int freed_tables : 1; struct mmu_gather_batch *active; struct mmu_gather_batch local; @@ -137,6 +147,7 @@ static inline void __tlb_reset_range(str tlb->start = TASK_SIZE; tlb->end = 0; } + tlb->freed_tables = 0; } static inline void tlb_flush_mmu_tlbonly(struct mmu_gather *tlb) @@ -278,6 +289,7 @@ static inline void tlb_remove_check_page #define pte_free_tlb(tlb, ptep, address) \ do { \ __tlb_adjust_range(tlb, address, PAGE_SIZE); \ + tlb->freed_tables = 1; \ __pte_free_tlb(tlb, ptep, address); \ } while (0) #endif @@ -285,7 +297,8 @@ static inline void tlb_remove_check_page #ifndef pmd_free_tlb #define pmd_free_tlb(tlb, pmdp, address) \ do { \ - __tlb_adjust_range(tlb, address, PAGE_SIZE); \ + __tlb_adjust_range(tlb, address, PAGE_SIZE); \ + tlb->freed_tables = 1; \ __pmd_free_tlb(tlb, pmdp, address); \ } while (0) #endif @@ -295,6 +308,7 @@ static inline void tlb_remove_check_page #define pud_free_tlb(tlb, pudp, address) \ do { \ __tlb_adjust_range(tlb, address, PAGE_SIZE); \ + tlb->freed_tables = 1; \ __pud_free_tlb(tlb, pudp, address); \ } while (0) #endif @@ -304,7 +318,8 @@ static inline void tlb_remove_check_page #ifndef p4d_free_tlb #define p4d_free_tlb(tlb, pudp, address) \ do { \ - __tlb_adjust_range(tlb, address, PAGE_SIZE); \ + __tlb_adjust_range(tlb, address, PAGE_SIZE); \ + tlb->freed_tables = 1; \ __p4d_free_tlb(tlb, pudp, address); \ } while (0) #endif Patches currently in stable-queue which might be from sant...@fossix.org are queue-4.19/asm-generic-tlb-track-which-levels-of-the-page-tables-have-been-cleared.patch queue-4.19/asm-generic-tlb-track-freeing-of-page-table-directories-in-struct-mmu_gather.patch queue-4.19/asm-generic-tlb-avoid-potential-double-flush.patch queue-4.19/mm-mmu_gather-invalidate-tlb-correctly-on-batch-allocation-failure-and-flush.patch queue-4.19/powerpc-mmu_gather-enable-rcu_table_free-even-for-smp-case.patch queue-4.19/asm-generic-tlb-arch-invert-config_have_rcu_table_invalidate.patch