Rename cpu_corgroup_{map,mask} to cpu_corgrp_{map,mask} to free up the cpu_corgroup_* namespace. cpu_corgroup_mask() will be added back in the subsequent commit for CONFIG_SCHED_MC enablement.
No functional changes intended. Signed-off-by: K Prateek Nayak <kprateek.na...@amd.com> --- arch/powerpc/kernel/smp.c | 26 +++++++++++++------------- 1 file changed, 13 insertions(+), 13 deletions(-) diff --git a/arch/powerpc/kernel/smp.c b/arch/powerpc/kernel/smp.c index 862f50c09539..4f48262658cc 100644 --- a/arch/powerpc/kernel/smp.c +++ b/arch/powerpc/kernel/smp.c @@ -87,7 +87,7 @@ DEFINE_PER_CPU(cpumask_var_t, cpu_sibling_map); DEFINE_PER_CPU(cpumask_var_t, cpu_smallcore_map); DEFINE_PER_CPU(cpumask_var_t, cpu_l2_cache_map); DEFINE_PER_CPU(cpumask_var_t, cpu_core_map); -static DEFINE_PER_CPU(cpumask_var_t, cpu_coregroup_map); +static DEFINE_PER_CPU(cpumask_var_t, cpu_corgrp_map); EXPORT_PER_CPU_SYMBOL(cpu_sibling_map); EXPORT_PER_CPU_SYMBOL(cpu_l2_cache_map); @@ -1045,9 +1045,9 @@ static const struct cpumask *tl_smt_mask(struct sched_domain_topology_level *tl, } #endif -static struct cpumask *cpu_coregroup_mask(int cpu) +static struct cpumask *cpu_corgrp_mask(int cpu) { - return per_cpu(cpu_coregroup_map, cpu); + return per_cpu(cpu_corgrp_map, cpu); } static bool has_coregroup_support(void) @@ -1061,7 +1061,7 @@ static bool has_coregroup_support(void) static const struct cpumask *cpu_mc_mask(struct sched_domain_topology_level *tl, int cpu) { - return cpu_coregroup_mask(cpu); + return cpu_corgrp_mask(cpu); } static const struct cpumask *cpu_pkg_mask(struct sched_domain_topology_level *tl, int cpu) @@ -1124,7 +1124,7 @@ void __init smp_prepare_cpus(unsigned int max_cpus) zalloc_cpumask_var_node(&per_cpu(cpu_core_map, cpu), GFP_KERNEL, cpu_to_node(cpu)); if (has_coregroup_support()) - zalloc_cpumask_var_node(&per_cpu(cpu_coregroup_map, cpu), + zalloc_cpumask_var_node(&per_cpu(cpu_corgrp_map, cpu), GFP_KERNEL, cpu_to_node(cpu)); #ifdef CONFIG_NUMA @@ -1145,7 +1145,7 @@ void __init smp_prepare_cpus(unsigned int max_cpus) cpumask_set_cpu(boot_cpuid, cpu_core_mask(boot_cpuid)); if (has_coregroup_support()) - cpumask_set_cpu(boot_cpuid, cpu_coregroup_mask(boot_cpuid)); + cpumask_set_cpu(boot_cpuid, cpu_corgrp_mask(boot_cpuid)); init_big_cores(); if (has_big_cores) { @@ -1510,8 +1510,8 @@ static void remove_cpu_from_masks(int cpu) set_cpus_unrelated(cpu, i, cpu_core_mask); if (has_coregroup_support()) { - for_each_cpu(i, cpu_coregroup_mask(cpu)) - set_cpus_unrelated(cpu, i, cpu_coregroup_mask); + for_each_cpu(i, cpu_corgrp_mask(cpu)) + set_cpus_unrelated(cpu, i, cpu_corgrp_mask); } } #endif @@ -1543,7 +1543,7 @@ static void update_coregroup_mask(int cpu, cpumask_var_t *mask) if (!*mask) { /* Assume only siblings are part of this CPU's coregroup */ for_each_cpu(i, submask_fn(cpu)) - set_cpus_related(cpu, i, cpu_coregroup_mask); + set_cpus_related(cpu, i, cpu_corgrp_mask); return; } @@ -1551,18 +1551,18 @@ static void update_coregroup_mask(int cpu, cpumask_var_t *mask) cpumask_and(*mask, cpu_online_mask, cpu_node_mask(cpu)); /* Update coregroup mask with all the CPUs that are part of submask */ - or_cpumasks_related(cpu, cpu, submask_fn, cpu_coregroup_mask); + or_cpumasks_related(cpu, cpu, submask_fn, cpu_corgrp_mask); /* Skip all CPUs already part of coregroup mask */ - cpumask_andnot(*mask, *mask, cpu_coregroup_mask(cpu)); + cpumask_andnot(*mask, *mask, cpu_corgrp_mask(cpu)); for_each_cpu(i, *mask) { /* Skip all CPUs not part of this coregroup */ if (coregroup_id == cpu_to_coregroup_id(i)) { - or_cpumasks_related(cpu, i, submask_fn, cpu_coregroup_mask); + or_cpumasks_related(cpu, i, submask_fn, cpu_corgrp_mask); cpumask_andnot(*mask, *mask, submask_fn(i)); } else { - cpumask_andnot(*mask, *mask, cpu_coregroup_mask(i)); + cpumask_andnot(*mask, *mask, cpu_corgrp_mask(i)); } } } -- 2.34.1