Le 26/08/2025 à 06:13, K Prateek Nayak a écrit :
Rename cpu_corgroup_{map,mask} to cpu_corgrp_{map,mask} to free up the
cpu_corgroup_* namespace. cpu_corgroup_mask() will be added back in the
subsequent commit for CONFIG_SCHED_MC enablement.

This renaming seems odd and uncomplete. For instance update_coregroup_mask() should probably be renamed as well shoudln't it ?

When you say cpu_corgroup_mask() will be added back, you mean the same function or a completely different function but with the same name ?

What's really the difference between corgrp and coregroup ?

Shouldn't also has_coregroup_support() now be renamed has_corgrp_support() ?

Christophe


No functional changes intended.

Signed-off-by: K Prateek Nayak <kprateek.na...@amd.com>
---
  arch/powerpc/kernel/smp.c | 26 +++++++++++++-------------
  1 file changed, 13 insertions(+), 13 deletions(-)

diff --git a/arch/powerpc/kernel/smp.c b/arch/powerpc/kernel/smp.c
index 862f50c09539..4f48262658cc 100644
--- a/arch/powerpc/kernel/smp.c
+++ b/arch/powerpc/kernel/smp.c
@@ -87,7 +87,7 @@ DEFINE_PER_CPU(cpumask_var_t, cpu_sibling_map);
  DEFINE_PER_CPU(cpumask_var_t, cpu_smallcore_map);
  DEFINE_PER_CPU(cpumask_var_t, cpu_l2_cache_map);
  DEFINE_PER_CPU(cpumask_var_t, cpu_core_map);
-static DEFINE_PER_CPU(cpumask_var_t, cpu_coregroup_map);
+static DEFINE_PER_CPU(cpumask_var_t, cpu_corgrp_map);
EXPORT_PER_CPU_SYMBOL(cpu_sibling_map);
  EXPORT_PER_CPU_SYMBOL(cpu_l2_cache_map);
@@ -1045,9 +1045,9 @@ static const struct cpumask *tl_smt_mask(struct 
sched_domain_topology_level *tl,
  }
  #endif
-static struct cpumask *cpu_coregroup_mask(int cpu)
+static struct cpumask *cpu_corgrp_mask(int cpu)
  {
-       return per_cpu(cpu_coregroup_map, cpu);
+       return per_cpu(cpu_corgrp_map, cpu);
  }
static bool has_coregroup_support(void)
@@ -1061,7 +1061,7 @@ static bool has_coregroup_support(void)
static const struct cpumask *cpu_mc_mask(struct sched_domain_topology_level *tl, int cpu)
  {
-       return cpu_coregroup_mask(cpu);
+       return cpu_corgrp_mask(cpu);
  }
static const struct cpumask *cpu_pkg_mask(struct sched_domain_topology_level *tl, int cpu)
@@ -1124,7 +1124,7 @@ void __init smp_prepare_cpus(unsigned int max_cpus)
                zalloc_cpumask_var_node(&per_cpu(cpu_core_map, cpu),
                                        GFP_KERNEL, cpu_to_node(cpu));
                if (has_coregroup_support())
-                       zalloc_cpumask_var_node(&per_cpu(cpu_coregroup_map, 
cpu),
+                       zalloc_cpumask_var_node(&per_cpu(cpu_corgrp_map, cpu),
                                                GFP_KERNEL, cpu_to_node(cpu));
#ifdef CONFIG_NUMA
@@ -1145,7 +1145,7 @@ void __init smp_prepare_cpus(unsigned int max_cpus)
        cpumask_set_cpu(boot_cpuid, cpu_core_mask(boot_cpuid));
if (has_coregroup_support())
-               cpumask_set_cpu(boot_cpuid, cpu_coregroup_mask(boot_cpuid));
+               cpumask_set_cpu(boot_cpuid, cpu_corgrp_mask(boot_cpuid));
init_big_cores();
        if (has_big_cores) {
@@ -1510,8 +1510,8 @@ static void remove_cpu_from_masks(int cpu)
                set_cpus_unrelated(cpu, i, cpu_core_mask);
if (has_coregroup_support()) {
-               for_each_cpu(i, cpu_coregroup_mask(cpu))
-                       set_cpus_unrelated(cpu, i, cpu_coregroup_mask);
+               for_each_cpu(i, cpu_corgrp_mask(cpu))
+                       set_cpus_unrelated(cpu, i, cpu_corgrp_mask);
        }
  }
  #endif
@@ -1543,7 +1543,7 @@ static void update_coregroup_mask(int cpu, cpumask_var_t 
*mask)
        if (!*mask) {
                /* Assume only siblings are part of this CPU's coregroup */
                for_each_cpu(i, submask_fn(cpu))
-                       set_cpus_related(cpu, i, cpu_coregroup_mask);
+                       set_cpus_related(cpu, i, cpu_corgrp_mask);
return;
        }
@@ -1551,18 +1551,18 @@ static void update_coregroup_mask(int cpu, 
cpumask_var_t *mask)
        cpumask_and(*mask, cpu_online_mask, cpu_node_mask(cpu));
/* Update coregroup mask with all the CPUs that are part of submask */
-       or_cpumasks_related(cpu, cpu, submask_fn, cpu_coregroup_mask);
+       or_cpumasks_related(cpu, cpu, submask_fn, cpu_corgrp_mask);
/* Skip all CPUs already part of coregroup mask */
-       cpumask_andnot(*mask, *mask, cpu_coregroup_mask(cpu));
+       cpumask_andnot(*mask, *mask, cpu_corgrp_mask(cpu));
for_each_cpu(i, *mask) {
                /* Skip all CPUs not part of this coregroup */
                if (coregroup_id == cpu_to_coregroup_id(i)) {
-                       or_cpumasks_related(cpu, i, submask_fn, 
cpu_coregroup_mask);
+                       or_cpumasks_related(cpu, i, submask_fn, 
cpu_corgrp_mask);
                        cpumask_andnot(*mask, *mask, submask_fn(i));
                } else {
-                       cpumask_andnot(*mask, *mask, cpu_coregroup_mask(i));
+                       cpumask_andnot(*mask, *mask, cpu_corgrp_mask(i));
                }
        }
  }


Reply via email to