Excerpts from Aneesh Kumar K.V's message of June 15, 2020 4:14 pm:
> This prepare kernel to operate with a different value than userspace AMR.
> For this, AMR needs to be saved and restored on entry and return from the
> kernel.
> 
> With KUAP we modify kernel AMR when accessing user address from the kernel
> via copy_to/from_user interfaces.
> 
> If MMU_FTR_KEY is enabled we always use the key mechanism to implement KUAP
> feature. If MMU_FTR_KEY is not supported and if we support MMU_FTR_KUAP
> (radix translation on POWER9), we can skip restoring AMR on return
> to userspace. Userspace won't be using AMR in that specific config.
> 
> Signed-off-by: Aneesh Kumar K.V <aneesh.ku...@linux.ibm.com>
> ---
>  arch/powerpc/include/asm/book3s/64/kup.h | 141 ++++++++++++++++++-----
>  arch/powerpc/kernel/entry_64.S           |   6 +-
>  arch/powerpc/kernel/exceptions-64s.S     |   4 +-
>  arch/powerpc/kernel/syscall_64.c         |  26 ++++-
>  4 files changed, 144 insertions(+), 33 deletions(-)
> 
> diff --git a/arch/powerpc/include/asm/book3s/64/kup.h 
> b/arch/powerpc/include/asm/book3s/64/kup.h
> index e6ee1c34842f..6979cd1a0003 100644
> --- a/arch/powerpc/include/asm/book3s/64/kup.h
> +++ b/arch/powerpc/include/asm/book3s/64/kup.h
> @@ -13,18 +13,47 @@
>  
>  #ifdef __ASSEMBLY__
>  
> -.macro kuap_restore_amr      gpr1, gpr2
> -#ifdef CONFIG_PPC_KUAP
> +.macro kuap_restore_user_amr gpr1
> +#if defined(CONFIG_PPC_MEM_KEYS)
>       BEGIN_MMU_FTR_SECTION_NESTED(67)
> -     mfspr   \gpr1, SPRN_AMR
> +     /*
> +      * AMR is going to be different when
> +      * returning to userspace.
> +      */
> +     ld      \gpr1, STACK_REGS_KUAP(r1)
> +     isync
> +     mtspr   SPRN_AMR, \gpr1
> +
> +     /* No isync required, see kuap_restore_user_amr() */
> +     END_MMU_FTR_SECTION_NESTED_IFSET(MMU_FTR_PKEY , 67)
> +#endif
> +.endm
> +
> +.macro kuap_restore_kernel_amr       gpr1, gpr2
> +#if defined(CONFIG_PPC_MEM_KEYS)
> +     BEGIN_MMU_FTR_SECTION_NESTED(67)
> +     b       99f  // handle_pkey_restore_amr
> +     END_MMU_FTR_SECTION_NESTED_IFSET(MMU_FTR_PKEY , 67)
> +
> +     BEGIN_MMU_FTR_SECTION_NESTED(68)
> +     b       99f  // handle_kuap_restore_amr
> +     MMU_FTR_SECTION_ELSE_NESTED(68)
> +     b       100f  // skip_restore_amr
> +     ALT_MMU_FTR_SECTION_END_NESTED_IFSET(MMU_FTR_KUAP, 68)
> +
> +99:
> +     /*
> +      * AMR is going to be mostly the same since we are
> +      * returning to the kernel. Compare and do a mtspr.
> +      */
>       ld      \gpr2, STACK_REGS_KUAP(r1)
> +     mfspr   \gpr1, SPRN_AMR
>       cmpd    \gpr1, \gpr2
> -     beq     998f
> +     beq     100f
>       isync
>       mtspr   SPRN_AMR, \gpr2
>       /* No isync required, see kuap_restore_amr() */
> -998:
> -     END_MMU_FTR_SECTION_NESTED_IFSET(MMU_FTR_KUAP, 67)
> +100:  // skip_restore_amr

Can't you code it like this? (_IFCLR requires none of the bits to be 
set)

BEGIN_MMU_FTR_SECTION_NESTED(67)
        b       99f     // nothing using AMR, no need to restore
END_MMU_FTR_SECTION_NESTED_IFCLR(MMU_FTR_PKEY | MMU_FTR_KUAP, 67)

That saves you a branch in the common case of using AMR. Similar
for others.

> @@ -69,22 +133,40 @@
>  
>  extern u64 default_uamor;
>  
> -static inline void kuap_restore_amr(struct pt_regs *regs, unsigned long amr)
> +static inline void kuap_restore_user_amr(struct pt_regs *regs)
>  {
> -     if (mmu_has_feature(MMU_FTR_KUAP) && unlikely(regs->kuap != amr)) {
> -             isync();
> -             mtspr(SPRN_AMR, regs->kuap);
> -             /*
> -              * No isync required here because we are about to RFI back to
> -              * previous context before any user accesses would be made,
> -              * which is a CSI.
> -              */
> +     if (!mmu_has_feature(MMU_FTR_PKEY))
> +             return;

If you have PKEY but not KUAP, do you still have to restore?

> +
> +     isync();
> +     mtspr(SPRN_AMR, regs->kuap);
> +     /*
> +      * No isync required here because we are about to rfi
> +      * back to previous context before any user accesses
> +      * would be made, which is a CSI.
> +      */
> +}
> +
> +static inline void kuap_restore_kernel_amr(struct pt_regs *regs,
> +                                        unsigned long amr)
> +{
> +     if (mmu_has_feature(MMU_FTR_KUAP) || mmu_has_feature(MMU_FTR_PKEY)) {
> +
> +             if (unlikely(regs->kuap != amr)) {
> +                     isync();
> +                     mtspr(SPRN_AMR, regs->kuap);
> +                     /*
> +                      * No isync required here because we are about to rfi
> +                      * back to previous context before any user accesses
> +                      * would be made, which is a CSI.
> +                      */
> +             }
>       }
>  }
>  
>  static inline unsigned long kuap_get_and_check_amr(void)
>  {
> -     if (mmu_has_feature(MMU_FTR_KUAP)) {
> +     if (mmu_has_feature(MMU_FTR_KUAP) || mmu_has_feature(MMU_FTR_PKEY)) {
>               unsigned long amr = mfspr(SPRN_AMR);
>               if (IS_ENABLED(CONFIG_PPC_KUAP_DEBUG)) /* kuap_check_amr() */
>                       WARN_ON_ONCE(amr != AMR_KUAP_BLOCKED);

We could do a static key that's based on this condition, but that can 
wait for another day.

>  
>  static inline void kuap_check_amr(void)
>  {
> -     if (IS_ENABLED(CONFIG_PPC_KUAP_DEBUG) && mmu_has_feature(MMU_FTR_KUAP))
> +     if (IS_ENABLED(CONFIG_PPC_KUAP_DEBUG) &&
> +         (mmu_has_feature(MMU_FTR_KUAP) || mmu_has_feature(MMU_FTR_PKEY)))
>               WARN_ON_ONCE(mfspr(SPRN_AMR) != AMR_KUAP_BLOCKED);
>  }
>  
>  #else /* CONFIG_PPC_MEM_KEYS */
>  
> -static inline void kuap_restore_amr(struct pt_regs *regs, unsigned long amr)
> +static inline void kuap_restore_user_amr(struct pt_regs *regs)
> +{
> +}
> +
> +static inline void kuap_restore_kernel_amr(struct pt_regs *regs, unsigned 
> long amr)
>  {
>  }
>  
> @@ -113,6 +200,7 @@ static inline unsigned long kuap_get_and_check_amr(void)
>  {
>       return 0;
>  }
> +
>  #endif /* CONFIG_PPC_MEM_KEYS */
>  
>  
> @@ -187,7 +275,6 @@ bad_kuap_fault(struct pt_regs *regs, unsigned long 
> address, bool is_write)
>                   "Bug: %s fault blocked by AMR!", is_write ? "Write" : 
> "Read");
>  }
>  #endif /* CONFIG_PPC_KUAP */
> -
>  #endif /* __ASSEMBLY__ */
>  
>  #endif /* _ASM_POWERPC_BOOK3S_64_KUP_H */

Unnecessary hunks.

> diff --git a/arch/powerpc/kernel/entry_64.S b/arch/powerpc/kernel/entry_64.S
> index 9d49338e0c85..a087cbe0b17d 100644
> --- a/arch/powerpc/kernel/entry_64.S
> +++ b/arch/powerpc/kernel/entry_64.S
> @@ -481,8 +481,8 @@ _ASM_NOKPROBE_SYMBOL(fast_interrupt_return)
>       kuap_check_amr r3, r4
>       ld      r5,_MSR(r1)
>       andi.   r0,r5,MSR_PR
> -     bne     .Lfast_user_interrupt_return
> -     kuap_restore_amr r3, r4
> +     bne     .Lfast_user_interrupt_return_amr
> +     kuap_restore_kernel_amr r3, r4
>       andi.   r0,r5,MSR_RI
>       li      r3,0 /* 0 return value, no EMULATE_STACK_STORE */
>       bne+    .Lfast_kernel_interrupt_return
> @@ -502,6 +502,8 @@ _ASM_NOKPROBE_SYMBOL(interrupt_return)
>       cmpdi   r3,0
>       bne-    .Lrestore_nvgprs
>  
> +.Lfast_user_interrupt_return_amr:
> +     kuap_restore_user_amr r3
>  .Lfast_user_interrupt_return:
>       ld      r11,_NIP(r1)
>       ld      r12,_MSR(r1)
> diff --git a/arch/powerpc/kernel/exceptions-64s.S 
> b/arch/powerpc/kernel/exceptions-64s.S
> index e70ebb5c318c..8226af444d77 100644
> --- a/arch/powerpc/kernel/exceptions-64s.S
> +++ b/arch/powerpc/kernel/exceptions-64s.S
> @@ -971,7 +971,7 @@ EXC_COMMON_BEGIN(system_reset_common)
>       ld      r10,SOFTE(r1)
>       stb     r10,PACAIRQSOFTMASK(r13)
>  
> -     kuap_restore_amr r9, r10
> +     kuap_restore_kernel_amr r9, r10
>       EXCEPTION_RESTORE_REGS
>       RFI_TO_USER_OR_KERNEL
>  
> @@ -2784,7 +2784,7 @@ EXC_COMMON_BEGIN(soft_nmi_common)
>       ld      r10,SOFTE(r1)
>       stb     r10,PACAIRQSOFTMASK(r13)
>  
> -     kuap_restore_amr r9, r10
> +     kuap_restore_kernel_amr r9, r10
>       EXCEPTION_RESTORE_REGS hsrr=0
>       RFI_TO_KERNEL
>  
> diff --git a/arch/powerpc/kernel/syscall_64.c 
> b/arch/powerpc/kernel/syscall_64.c
> index 7e560a01afa4..fded67982fbe 100644
> --- a/arch/powerpc/kernel/syscall_64.c
> +++ b/arch/powerpc/kernel/syscall_64.c
> @@ -35,7 +35,21 @@ notrace long system_call_exception(long r3, long r4, long 
> r5,
>       BUG_ON(!FULL_REGS(regs));
>       BUG_ON(regs->softe != IRQS_ENABLED);
>  
> -     kuap_check_amr();
> +#ifdef CONFIG_PPC_MEM_KEYS
> +     if (mmu_has_feature(MMU_FTR_PKEY)) {
> +             unsigned long amr;
> +             /*
> +              * When entering from userspace we mostly have the AMR
> +              * different from kernel default values. Hence don't compare.
> +              */
> +             amr = mfspr(SPRN_AMR);
> +             regs->kuap = amr;
> +             if (mmu_has_feature(MMU_FTR_KUAP))
> +                     mtspr(SPRN_AMR, AMR_KUAP_BLOCKED);
> +             isync();

isync should be inside the if(). Again do pkeys need to save this if 
KUAP is not being used? I haven't really looked at how all that works, 
but what's changing for the PKEY && !KUAP case?

This would be nice if it could all go into a wrapper function rather 
than ifdef.

> +     } else
> +#endif
> +             kuap_check_amr();
>  
>       account_cpu_user_entry();
>  
> @@ -222,6 +236,10 @@ notrace unsigned long syscall_exit_prepare(unsigned long 
> r3,
>  
>       account_cpu_user_exit();
>  
> +     /*
> +      * We do this at the end so that we do context switch with KERNEL AMR
> +      */
> +     kuap_restore_user_amr(regs);
>       return ret;

Comment doesn't make sense, newline required before return.

>  }
>  
> @@ -306,6 +324,10 @@ notrace unsigned long interrupt_exit_user_prepare(struct 
> pt_regs *regs, unsigned
>  
>       account_cpu_user_exit();
>  
> +     /*
> +      * We do this at the end so that we do context switch with KERNEL AMR
> +      */
> +     kuap_restore_user_amr(regs);

Duplicated comments I prefer to just have like this instead of trying to 
keep them in sync. Can complete the circular reference by having a

* similarly in interrupt_exit_user_prepare

in the main comment, but if they come close to one another in the same 
file it's not so important to keep them together.

 +      kuap_restore_user_amr(regs); /* see syscall_exit_prepare */

>       return ret;
>  }
>  
> @@ -376,7 +398,7 @@ notrace unsigned long 
> interrupt_exit_kernel_prepare(struct pt_regs *regs, unsign
>        * which would cause Read-After-Write stalls. Hence, we take the AMR
>        * value from the check above.
>        */
> -     kuap_restore_amr(regs, amr);
> +     kuap_restore_kernel_amr(regs, amr);
>  
>       return ret;
>  }
> -- 
> 2.26.2
> 
> 

Reply via email to