The branch main has been updated by kib:

URL: 
https://cgit.FreeBSD.org/src/commit/?id=120ca8d74b46caa260702485e30fe5f9f9984682

commit 120ca8d74b46caa260702485e30fe5f9f9984682
Author:     Konstantin Belousov <[email protected]>
AuthorDate: 2026-02-02 20:22:46 +0000
Commit:     Konstantin Belousov <[email protected]>
CommitDate: 2026-02-03 02:43:18 +0000

    Re-introduce kern.sched.topology_spec
    
    Move it back from kern.sched.ule.topology_spec.
    Make it scheduler-agnostic.
    Provide trivial report for UP kernels.
    
    Apparently the MIB is used by some third-party software.  Obviously it
    did not worked on UP or 4BSD configs.
    
    PR:     292574
    Reviewed by:    olce
    Sponsored by:   The FreeBSD Foundation
    MFC after:      1 week
    Differential revision:  https://reviews.freebsd.org/D55062
---
 sys/kern/sched_shim.c | 91 ++++++++++++++++++++++++++++++++++++++++++++++++++
 sys/kern/sched_ule.c  | 92 ---------------------------------------------------
 sys/kern/subr_smp.c   | 81 ++++++++++++++++++++++++++-------------------
 sys/sys/smp.h         |  9 +++--
 4 files changed, 144 insertions(+), 129 deletions(-)

diff --git a/sys/kern/sched_shim.c b/sys/kern/sched_shim.c
index ec5c42c37aab..83e4412494d3 100644
--- a/sys/kern/sched_shim.c
+++ b/sys/kern/sched_shim.c
@@ -16,6 +16,7 @@
 #include <sys/runq.h>
 #include <sys/sbuf.h>
 #include <sys/sched.h>
+#include <sys/smp.h>
 #include <sys/sysctl.h>
 #include <machine/ifunc.h>
 
@@ -171,9 +172,12 @@ schedinit(void)
        active_sched->init();
 }
 
+struct cpu_group __read_mostly *cpu_top;               /* CPU topology */
+
 static void
 sched_setup(void *dummy)
 {
+       cpu_top = smp_topo();
        active_sched->setup();
 }
 SYSINIT(sched_setup, SI_SUB_RUN_QUEUE, SI_ORDER_FIRST, sched_setup, NULL);
@@ -232,3 +236,90 @@ SYSCTL_PROC(_kern_sched, OID_AUTO, available,
 fixpt_t ccpu;
 SYSCTL_UINT(_kern, OID_AUTO, ccpu, CTLFLAG_RD, &ccpu, 0,
     "Decay factor used for updating %CPU");
+
+/*
+ * Build the CPU topology dump string. Is recursively called to collect
+ * the topology tree.
+ */
+static int
+sysctl_kern_sched_topology_spec_internal(struct sbuf *sb,
+    struct cpu_group *cg, int indent)
+{
+       char cpusetbuf[CPUSETBUFSIZ];
+       int i, first;
+
+       if (cpu_top == NULL) {
+               sbuf_printf(sb, "%*s<group level=\"1\" cache-level=\"1\">\n",
+                   indent, "");
+               sbuf_printf(sb, "%*s</group>\n", indent, "");
+               return (0);
+       }
+
+       sbuf_printf(sb, "%*s<group level=\"%d\" cache-level=\"%d\">\n", indent,
+           "", 1 + indent / 2, cg->cg_level);
+       sbuf_printf(sb, "%*s <cpu count=\"%d\" mask=\"%s\">", indent, "",
+           cg->cg_count, cpusetobj_strprint(cpusetbuf, &cg->cg_mask));
+       first = TRUE;
+       for (i = cg->cg_first; i <= cg->cg_last; i++) {
+               if (CPU_ISSET(i, &cg->cg_mask)) {
+                       if (!first)
+                               sbuf_cat(sb, ", ");
+                       else
+                               first = FALSE;
+                       sbuf_printf(sb, "%d", i);
+               }
+       }
+       sbuf_cat(sb, "</cpu>\n");
+
+       if (cg->cg_flags != 0) {
+               sbuf_printf(sb, "%*s <flags>", indent, "");
+               if ((cg->cg_flags & CG_FLAG_HTT) != 0)
+                       sbuf_cat(sb, "<flag name=\"HTT\">HTT group</flag>");
+               if ((cg->cg_flags & CG_FLAG_THREAD) != 0)
+                       sbuf_cat(sb, "<flag name=\"THREAD\">THREAD 
group</flag>");
+               if ((cg->cg_flags & CG_FLAG_SMT) != 0)
+                       sbuf_cat(sb, "<flag name=\"SMT\">SMT group</flag>");
+               if ((cg->cg_flags & CG_FLAG_NODE) != 0)
+                       sbuf_cat(sb, "<flag name=\"NODE\">NUMA node</flag>");
+               sbuf_cat(sb, "</flags>\n");
+       }
+
+       if (cg->cg_children > 0) {
+               sbuf_printf(sb, "%*s <children>\n", indent, "");
+               for (i = 0; i < cg->cg_children; i++)
+                       sysctl_kern_sched_topology_spec_internal(sb,
+                           &cg->cg_child[i], indent + 2);
+               sbuf_printf(sb, "%*s </children>\n", indent, "");
+       }
+       sbuf_printf(sb, "%*s</group>\n", indent, "");
+       return (0);
+}
+
+/*
+ * Sysctl handler for retrieving topology dump. It's a wrapper for
+ * the recursive sysctl_kern_smp_topology_spec_internal().
+ */
+static int
+sysctl_kern_sched_topology_spec(SYSCTL_HANDLER_ARGS)
+{
+       struct sbuf *topo;
+       int err;
+
+       topo = sbuf_new_for_sysctl(NULL, NULL, 512, req);
+       if (topo == NULL)
+               return (ENOMEM);
+
+       sbuf_cat(topo, "<groups>\n");
+       err = sysctl_kern_sched_topology_spec_internal(topo, cpu_top, 1);
+       sbuf_cat(topo, "</groups>\n");
+
+       if (err == 0)
+               err = sbuf_finish(topo);
+       sbuf_delete(topo);
+       return (err);
+}
+
+SYSCTL_PROC(_kern_sched, OID_AUTO, topology_spec, CTLTYPE_STRING |
+    CTLFLAG_MPSAFE | CTLFLAG_RD, NULL, 0,
+    sysctl_kern_sched_topology_spec, "A",
+    "XML dump of detected CPU topology");
diff --git a/sys/kern/sched_ule.c b/sys/kern/sched_ule.c
index ccad7947c4f5..c6bfe15e768b 100644
--- a/sys/kern/sched_ule.c
+++ b/sys/kern/sched_ule.c
@@ -304,7 +304,6 @@ struct tdq {
                                 atomic_load_short(&(tdq)->tdq_switchcnt) + 1))
 
 #ifdef SMP
-struct cpu_group __read_mostly *cpu_top;               /* CPU topology */
 
 #define        SCHED_AFFINITY_DEFAULT  (max(1, hz / 1000))
 /*
@@ -398,9 +397,6 @@ static void sched_balance(void);
 static bool sched_balance_pair(struct tdq *, struct tdq *);
 static inline struct tdq *sched_setcpu(struct thread *, int, int);
 static inline void thread_unblock_switch(struct thread *, struct mtx *);
-static int sysctl_kern_sched_ule_topology_spec(SYSCTL_HANDLER_ARGS);
-static int sysctl_kern_sched_ule_topology_spec_internal(struct sbuf *sb,
-    struct cpu_group *cg, int indent);
 #endif
 
 /*
@@ -1590,7 +1586,6 @@ sched_setup_smp(void)
        struct tdq *tdq;
        int i;
 
-       cpu_top = smp_topo();
        CPU_FOREACH(i) {
                tdq = DPCPU_ID_PTR(i, tdq);
                tdq_setup(tdq, i);
@@ -3452,89 +3447,6 @@ struct sched_instance sched_ule_instance = {
 };
 DECLARE_SCHEDULER(ule_sched_selector, "ULE", &sched_ule_instance);
 
-#ifdef SMP
-
-/*
- * Build the CPU topology dump string. Is recursively called to collect
- * the topology tree.
- */
-static int
-sysctl_kern_sched_ule_topology_spec_internal(struct sbuf *sb,
-    struct cpu_group *cg, int indent)
-{
-       char cpusetbuf[CPUSETBUFSIZ];
-       int i, first;
-
-       sbuf_printf(sb, "%*s<group level=\"%d\" cache-level=\"%d\">\n", indent,
-           "", 1 + indent / 2, cg->cg_level);
-       sbuf_printf(sb, "%*s <cpu count=\"%d\" mask=\"%s\">", indent, "",
-           cg->cg_count, cpusetobj_strprint(cpusetbuf, &cg->cg_mask));
-       first = TRUE;
-       for (i = cg->cg_first; i <= cg->cg_last; i++) {
-               if (CPU_ISSET(i, &cg->cg_mask)) {
-                       if (!first)
-                               sbuf_cat(sb, ", ");
-                       else
-                               first = FALSE;
-                       sbuf_printf(sb, "%d", i);
-               }
-       }
-       sbuf_cat(sb, "</cpu>\n");
-
-       if (cg->cg_flags != 0) {
-               sbuf_printf(sb, "%*s <flags>", indent, "");
-               if ((cg->cg_flags & CG_FLAG_HTT) != 0)
-                       sbuf_cat(sb, "<flag name=\"HTT\">HTT group</flag>");
-               if ((cg->cg_flags & CG_FLAG_THREAD) != 0)
-                       sbuf_cat(sb, "<flag name=\"THREAD\">THREAD 
group</flag>");
-               if ((cg->cg_flags & CG_FLAG_SMT) != 0)
-                       sbuf_cat(sb, "<flag name=\"SMT\">SMT group</flag>");
-               if ((cg->cg_flags & CG_FLAG_NODE) != 0)
-                       sbuf_cat(sb, "<flag name=\"NODE\">NUMA node</flag>");
-               sbuf_cat(sb, "</flags>\n");
-       }
-
-       if (cg->cg_children > 0) {
-               sbuf_printf(sb, "%*s <children>\n", indent, "");
-               for (i = 0; i < cg->cg_children; i++)
-                       sysctl_kern_sched_ule_topology_spec_internal(sb,
-                           &cg->cg_child[i], indent+2);
-               sbuf_printf(sb, "%*s </children>\n", indent, "");
-       }
-       sbuf_printf(sb, "%*s</group>\n", indent, "");
-       return (0);
-}
-
-/*
- * Sysctl handler for retrieving topology dump. It's a wrapper for
- * the recursive sysctl_kern_smp_topology_spec_internal().
- */
-static int
-sysctl_kern_sched_ule_topology_spec(SYSCTL_HANDLER_ARGS)
-{
-       struct sbuf *topo;
-       int err;
-
-       if (cpu_top == NULL)
-               return (ENOTTY);
-
-       topo = sbuf_new_for_sysctl(NULL, NULL, 512, req);
-       if (topo == NULL)
-               return (ENOMEM);
-
-       sbuf_cat(topo, "<groups>\n");
-       err = sysctl_kern_sched_ule_topology_spec_internal(topo, cpu_top, 1);
-       sbuf_cat(topo, "</groups>\n");
-
-       if (err == 0) {
-               err = sbuf_finish(topo);
-       }
-       sbuf_delete(topo);
-       return (err);
-}
-
-#endif
-
 static int
 sysctl_kern_quantum(SYSCTL_HANDLER_ARGS)
 {
@@ -3597,8 +3509,4 @@ SYSCTL_INT(_kern_sched_ule, OID_AUTO, trysteal_limit, 
CTLFLAG_RWTUN,
 SYSCTL_INT(_kern_sched_ule, OID_AUTO, always_steal, CTLFLAG_RWTUN,
     &always_steal, 0,
     "Always run the stealer from the idle thread");
-SYSCTL_PROC(_kern_sched_ule, OID_AUTO, topology_spec, CTLTYPE_STRING |
-    CTLFLAG_MPSAFE | CTLFLAG_RD, NULL, 0,
-    sysctl_kern_sched_ule_topology_spec, "A",
-    "XML dump of detected CPU topology");
 #endif
diff --git a/sys/kern/subr_smp.c b/sys/kern/subr_smp.c
index 353a69435971..2256ba648e4d 100644
--- a/sys/kern/subr_smp.c
+++ b/sys/kern/subr_smp.c
@@ -50,9 +50,43 @@
 
 #include "opt_sched.h"
 
-#ifdef SMP
 MALLOC_DEFINE(M_TOPO, "toponodes", "SMP topology data");
 
+struct cpu_group *
+smp_topo_alloc(u_int count)
+{
+       static struct cpu_group *group = NULL;
+       static u_int index;
+       u_int curr;
+
+       if (group == NULL) {
+               group = mallocarray((mp_maxid + 1) * MAX_CACHE_LEVELS + 1,
+                   sizeof(*group), M_DEVBUF, M_WAITOK | M_ZERO);
+       }
+       curr = index;
+       index += count;
+       return (&group[curr]);
+}
+
+struct cpu_group *
+smp_topo_none(void)
+{
+       struct cpu_group *top;
+
+       top = smp_topo_alloc(1);
+       top->cg_parent = NULL;
+       top->cg_child = NULL;
+       top->cg_mask = all_cpus;
+       top->cg_count = mp_ncpus;
+       top->cg_children = 0;
+       top->cg_level = CG_SHARE_NONE;
+       top->cg_flags = 0;
+
+       return (top);
+}
+
+#ifdef SMP
+
 volatile cpuset_t stopped_cpus;
 volatile cpuset_t started_cpus;
 volatile cpuset_t suspended_cpus;
@@ -731,39 +765,6 @@ smp_topo(void)
        return (top);
 }
 
-struct cpu_group *
-smp_topo_alloc(u_int count)
-{
-       static struct cpu_group *group = NULL;
-       static u_int index;
-       u_int curr;
-
-       if (group == NULL) {
-               group = mallocarray((mp_maxid + 1) * MAX_CACHE_LEVELS + 1,
-                   sizeof(*group), M_DEVBUF, M_WAITOK | M_ZERO);
-       }
-       curr = index;
-       index += count;
-       return (&group[curr]);
-}
-
-struct cpu_group *
-smp_topo_none(void)
-{
-       struct cpu_group *top;
-
-       top = smp_topo_alloc(1);
-       top->cg_parent = NULL;
-       top->cg_child = NULL;
-       top->cg_mask = all_cpus;
-       top->cg_count = mp_ncpus;
-       top->cg_children = 0;
-       top->cg_level = CG_SHARE_NONE;
-       top->cg_flags = 0;
-
-       return (top);
-}
-
 static int
 smp_topo_addleaf(struct cpu_group *parent, struct cpu_group *child, int share,
     int count, int flags, int start)
@@ -901,6 +902,18 @@ smp_rendezvous(void (*setup_func)(void *),
            arg);
 }
 
+struct cpu_group *
+smp_topo(void)
+{
+       static struct cpu_group *top = NULL;
+
+       if (top != NULL)
+               return (top);
+
+       top = smp_topo_none();
+       return (top);
+}
+
 /*
  * Provide dummy SMP support for UP kernels.  Modules that need to use SMP
  * APIs will still work using this dummy support.
diff --git a/sys/sys/smp.h b/sys/sys/smp.h
index fdb69b13c0d4..493dc91043bd 100644
--- a/sys/sys/smp.h
+++ b/sys/sys/smp.h
@@ -89,6 +89,8 @@ struct cpu_group {
 
 typedef struct cpu_group *cpu_group_t;
 
+extern cpu_group_t cpu_top;
+
 /*
  * Defines common resources for CPUs in the group.  The highest level
  * resource should be used when multiple are shared.
@@ -147,9 +149,6 @@ int topo_analyze(struct topo_node *topo_root, int all,
 #define        TOPO_FOREACH(i, root)   \
        for (i = root; i != NULL; i = topo_next_node(root, i))
 
-struct cpu_group *smp_topo(void);
-struct cpu_group *smp_topo_alloc(u_int count);
-struct cpu_group *smp_topo_none(void);
 struct cpu_group *smp_topo_1level(int l1share, int l1count, int l1flags);
 struct cpu_group *smp_topo_2level(int l2share, int l2count, int l1share,
     int l1count, int l1flags);
@@ -166,6 +165,10 @@ extern cpuset_t hlt_cpus_mask;             /* XXX 'mask' 
is detail in old impl */
 extern cpuset_t logical_cpus_mask;
 #endif /* SMP */
 
+struct cpu_group *smp_topo(void);
+struct cpu_group *smp_topo_alloc(u_int count);
+struct cpu_group *smp_topo_none(void);
+
 extern u_int mp_maxid;
 extern int mp_maxcpus;
 extern int mp_ncores;

Reply via email to