merged.

Bruce


In message: [linux-yocto][v5.2/standard/base v5.2/standard/intel-x86][PATCH] 
Revert "x86: Use CONFIG_PREEMPTION"
on 16/01/2020 [email protected] wrote:

> From: He Zhe <[email protected]>
> 
> This reverts commit 41bbdde13b435400816397a47e84fa697934c253.
> 
> The reverted patch causes the following build failure.
> tmp-glibc/work-shared/intel-x86-64/kernel-source/include/linux/preempt.h:190:3:
> error: implicit declaration of function '__preempt_schedule'
> [-Werror=implicit-function-declaration]
> 
> The reverted patch should have been used with b8d3349803ba
> ("sched/rt, Kconfig: Unbreak def/oldconfig with CONFIG_PREEMPT=y") which fixes
> a50a3f4b6a31 ("sched/rt, Kconfig: Introduce CONFIG_PREEMPT_RT") which is not
> included in 5.2.29. So we simply revert 41bbdde13b43.
> 
> Signed-off-by: He Zhe <[email protected]>
> ---
>  arch/x86/entry/entry_32.S      | 6 +++---
>  arch/x86/entry/entry_64.S      | 4 ++--
>  arch/x86/entry/thunk_32.S      | 2 +-
>  arch/x86/entry/thunk_64.S      | 4 ++--
>  arch/x86/include/asm/preempt.h | 2 +-
>  arch/x86/kernel/kprobes/core.c | 2 +-
>  6 files changed, 10 insertions(+), 10 deletions(-)
> 
> diff --git a/arch/x86/entry/entry_32.S b/arch/x86/entry/entry_32.S
> index f07baf0388bc..1153e510cedd 100644
> --- a/arch/x86/entry/entry_32.S
> +++ b/arch/x86/entry/entry_32.S
> @@ -63,7 +63,7 @@
>   * enough to patch inline, increasing performance.
>   */
>  
> -#ifdef CONFIG_PREEMPTION
> +#ifdef CONFIG_PREEMPT
>  # define preempt_stop(clobbers)      DISABLE_INTERRUPTS(clobbers); 
> TRACE_IRQS_OFF
>  #else
>  # define preempt_stop(clobbers)
> @@ -1104,7 +1104,7 @@ restore_all:
>       INTERRUPT_RETURN
>  
>  restore_all_kernel:
> -#ifdef CONFIG_PREEMPTION
> +#ifdef CONFIG_PREEMPT
>       DISABLE_INTERRUPTS(CLBR_ANY)
>       cmpl    $0, PER_CPU_VAR(__preempt_count)
>       jnz     .Lno_preempt
> @@ -1393,7 +1393,7 @@ ENTRY(xen_hypervisor_callback)
>       TRACE_IRQS_OFF
>       mov     %esp, %eax
>       call    xen_evtchn_do_upcall
> -#ifndef CONFIG_PREEMPTION
> +#ifndef CONFIG_PREEMPT
>       call    xen_maybe_preempt_hcall
>  #endif
>       jmp     ret_from_intr
> diff --git a/arch/x86/entry/entry_64.S b/arch/x86/entry/entry_64.S
> index c2604cb55c63..7d81e14b6226 100644
> --- a/arch/x86/entry/entry_64.S
> +++ b/arch/x86/entry/entry_64.S
> @@ -664,7 +664,7 @@ GLOBAL(swapgs_restore_regs_and_return_to_usermode)
>  
>  /* Returning to kernel space */
>  retint_kernel:
> -#ifdef CONFIG_PREEMPTION
> +#ifdef CONFIG_PREEMPT
>       /* Interrupts are off */
>       /* Check if we need preemption */
>       btl     $9, EFLAGS(%rsp)                /* were interrupts off? */
> @@ -1115,7 +1115,7 @@ ENTRY(xen_do_hypervisor_callback)               /* 
> do_hypervisor_callback(struct *pt_regs) */
>       call    xen_evtchn_do_upcall
>       LEAVE_IRQ_STACK
>  
> -#ifndef CONFIG_PREEMPTION
> +#ifndef CONFIG_PREEMPT
>       call    xen_maybe_preempt_hcall
>  #endif
>       jmp     error_exit
> diff --git a/arch/x86/entry/thunk_32.S b/arch/x86/entry/thunk_32.S
> index 2713490611a3..cb3464525b37 100644
> --- a/arch/x86/entry/thunk_32.S
> +++ b/arch/x86/entry/thunk_32.S
> @@ -34,7 +34,7 @@
>       THUNK trace_hardirqs_off_thunk,trace_hardirqs_off_caller,1
>  #endif
>  
> -#ifdef CONFIG_PREEMPTION
> +#ifdef CONFIG_PREEMPT
>       THUNK ___preempt_schedule, preempt_schedule
>       THUNK ___preempt_schedule_notrace, preempt_schedule_notrace
>       EXPORT_SYMBOL(___preempt_schedule)
> diff --git a/arch/x86/entry/thunk_64.S b/arch/x86/entry/thunk_64.S
> index d6d1fdc09f9e..cfdca8b42c70 100644
> --- a/arch/x86/entry/thunk_64.S
> +++ b/arch/x86/entry/thunk_64.S
> @@ -47,7 +47,7 @@
>       THUNK lockdep_sys_exit_thunk,lockdep_sys_exit
>  #endif
>  
> -#ifdef CONFIG_PREEMPTION
> +#ifdef CONFIG_PREEMPT
>       THUNK ___preempt_schedule, preempt_schedule
>       THUNK ___preempt_schedule_notrace, preempt_schedule_notrace
>       EXPORT_SYMBOL(___preempt_schedule)
> @@ -56,7 +56,7 @@
>  
>  #if defined(CONFIG_TRACE_IRQFLAGS) \
>   || defined(CONFIG_DEBUG_LOCK_ALLOC) \
> - || defined(CONFIG_PREEMPTION)
> + || defined(CONFIG_PREEMPT)
>  .L_restore:
>       popq %r11
>       popq %r10
> diff --git a/arch/x86/include/asm/preempt.h b/arch/x86/include/asm/preempt.h
> index 3d4cb83a8828..99a7fa9ab0a3 100644
> --- a/arch/x86/include/asm/preempt.h
> +++ b/arch/x86/include/asm/preempt.h
> @@ -102,7 +102,7 @@ static __always_inline bool should_resched(int 
> preempt_offset)
>       return unlikely(raw_cpu_read_4(__preempt_count) == preempt_offset);
>  }
>  
> -#ifdef CONFIG_PREEMPTION
> +#ifdef CONFIG_PREEMPT
>    extern asmlinkage void ___preempt_schedule(void);
>  # define __preempt_schedule() \
>       asm volatile ("call ___preempt_schedule" : ASM_CALL_CONSTRAINT)
> diff --git a/arch/x86/kernel/kprobes/core.c b/arch/x86/kernel/kprobes/core.c
> index c88c1ec5938a..bd17dbb15d6a 100644
> --- a/arch/x86/kernel/kprobes/core.c
> +++ b/arch/x86/kernel/kprobes/core.c
> @@ -580,7 +580,7 @@ static void setup_singlestep(struct kprobe *p, struct 
> pt_regs *regs,
>       if (setup_detour_execution(p, regs, reenter))
>               return;
>  
> -#if !defined(CONFIG_PREEMPTION)
> +#if !defined(CONFIG_PREEMPT)
>       if (p->ainsn.boostable && !p->post_handler) {
>               /* Boost up -- we can execute copied instructions directly */
>               if (!reenter)
> -- 
> 2.24.1
> 
-=-=-=-=-=-=-=-=-=-=-=-
Links: You receive all messages sent to this group.

View/Reply Online (#8324): 
https://lists.yoctoproject.org/g/linux-yocto/message/8324
Mute This Topic: https://lists.yoctoproject.org/mt/69742345/21656
Group Owner: [email protected]
Unsubscribe: https://lists.yoctoproject.org/g/linux-yocto/unsub  
[[email protected]]
-=-=-=-=-=-=-=-=-=-=-=-

Reply via email to