On Thu, 6 Sep 2018, Jiri Kosina wrote:
> +/*
> + * The read-modify-write of the MSR doesn't need any race protection here,
> + * as we're running in atomic context.
> + */
> +static void enable_stibp(void *info)
> +{
> +     u64 mask;
> +     rdmsrl(MSR_IA32_SPEC_CTRL, mask);
> +     mask |= SPEC_CTRL_STIBP;
> +     wrmsrl(MSR_IA32_SPEC_CTRL, mask);
> +}

You need to write x86_spec_ctrl_base and you have to set/clear the STIBP
bit before issuing the IPIs, i.e. you only need a single function

static void update_specctrl_msr(void *unused)
{
        wrmsrl(MSR_IA32_SPEC_CTRL, x86_spec_ctrl_base);
}

If you don't update x86_spec_ctrl_base then any consecutive update of the
MSR will clear it. It's written in various nospec() functions and on
vmexit.

Right now the variable is ro_after_init because it's only modified during
init when the mitigations are selected. So that needs to be removed.

For the case at hand we could just rely on the hotplug serialization, but I
fear that's going to break sooner than later. I rather prefer to have some
local protection here right away.

void arch_smt_control(bool enable)
{
        mutex_lock(&spec_ctrl_mutex);
        if (enable)
                x86_spec_ctrl_base |= SPEC_CTRL_STIBP;
        else
                x86_spec_ctrl_base &= ~SPEC_CTRL_STIBP;

        on_each_cpu(update_specctrl_msr, NULL, 1);
        mutex_unlock(&spec_ctrl_mutex);
}

or something like that.

>  #undef pr_fmt
> @@ -655,6 +708,16 @@ void x86_spec_ctrl_setup_ap(void)
>  
>       if (ssb_mode == SPEC_STORE_BYPASS_DISABLE)
>               x86_amd_ssb_disable();
> +
> +     /*
> +      * If we are here during system bootup, enable STIBP.
> +      *
> +      * If we are here because of SMT hotplug, STIBP will be enabled by the
> +      * SMT control code (enabling here would not be sufficient, as it
> +      * needs to happen on primary threads as well).
> +      */
> +     if (stibp_needed() && system_state < SYSTEM_RUNNING)
> +             enable_stibp(NULL);

That's broken and pointless. If a CPU is hotplugged for whatever reason,
then it needs the update of the IA32_SPEC_CTRL msr and it's already there:

void x86_spec_ctrl_setup_ap(void)
{
        if (boot_cpu_has(X86_FEATURE_MSR_SPEC_CTRL))
                wrmsrl(MSR_IA32_SPEC_CTRL, x86_spec_ctrl_base);

so again. x86_spec_ctrl_base needs the STIPB bit update when the
enable/disable changes.

> +/*
> + * Architectures that need SMT-specific errata handling during SMT hotplug
> + * should override these.
> + */
> +void __weak arch_smt_enable_errata(void) { };
> +void __weak arch_smt_disable_errata(void) { };
> +
>  static int cpuhp_smt_disable(enum cpuhp_smt_control ctrlval)
>  {
>       int cpu, ret = 0;
>  
>       cpu_maps_update_begin();
> +     arch_smt_disable_errata();
>       for_each_online_cpu(cpu) {
>               if (topology_is_primary_thread(cpu))
>                       continue;
>               ret = cpu_down_maps_locked(cpu, CPUHP_OFFLINE);
> -             if (ret)
> +             if (ret) {
> +                     arch_smt_enable_errata();
>                       break;
> +             }

Why don't you do the disable_errata call _after_ the loop and only on
success? That's more logical and spares 50% IPIs

> @@ -2073,6 +2083,7 @@ static int cpuhp_smt_enable(void)
>                /* See comment in cpuhp_smt_disable() */
>                cpuhp_online_cpu_device(cpu);
>        }
> +       arch_smt_enable_errata();
>         cpu_maps_update_done();
>        return ret;

If you do that _before_ the loop then you spare 50% IPIs again because the
siblings will do the right thing via x86_spec_ctrl_base.

Thanks,

        tglx

Reply via email to