gpu.compute.weight
      A read-write flat-keyed file which exists on all cgroups.  The
      default weight is 100.  Each entry is keyed by the DRM device's
      major:minor (the primary minor).  The weights are in the range [1,
      10000] and specifies the relative amount of physical partitions
      the cgroup can use in relation to its siblings.  The partition
      concept here is analogous to the subdevice of OpenCL.

gpu.compute.effective
      A read-only nested-keyed file which exists on all cgroups.  Each
      entry is keyed by the DRM device's major:minor.

      It lists the GPU subdevices that are actually granted to this
      cgroup by its parent.  These subdevices are allowed to be used by
      tasks within the current cgroup.

      =====     ==============================================
      count     The total number of granted subdevices
      list      Enumeration of the subdevices
      =====     ==============================================

Change-Id: Idde0ef9a331fd67bb9c7eb8ef9978439e6452488
Signed-off-by: Kenny Ho <kenny...@amd.com>
---
 Documentation/admin-guide/cgroup-v2.rst |  21 +++
 include/drm/drm_cgroup.h                |   3 +
 include/linux/cgroup_drm.h              |  16 +++
 kernel/cgroup/drm.c                     | 177 +++++++++++++++++++++++-
 4 files changed, 215 insertions(+), 2 deletions(-)

diff --git a/Documentation/admin-guide/cgroup-v2.rst 
b/Documentation/admin-guide/cgroup-v2.rst
index 581343472651..f92f1f4a64d4 100644
--- a/Documentation/admin-guide/cgroup-v2.rst
+++ b/Documentation/admin-guide/cgroup-v2.rst
@@ -2126,6 +2126,27 @@ GPU Interface Files
        Set largest allocation for /dev/dri/card1 to 4MB
        echo "226:1 4m" > gpu.buffer.peak.max
 
+  gpu.compute.weight
+       A read-write flat-keyed file which exists on all cgroups.  The
+       default weight is 100.  Each entry is keyed by the DRM device's
+       major:minor (the primary minor).  The weights are in the range
+       [1, 10000] and specifies the relative amount of physical partitions 
+       the cgroup can use in relation to its siblings.  The partition
+       concept here is analogous to the subdevice concept of OpenCL.
+
+  gpu.compute.effective
+       A read-only nested-keyed file which exists on all cgroups.
+       Each entry is keyed by the DRM device's major:minor.
+
+       It lists the GPU subdevices that are actually granted to this
+       cgroup by its parent.  These subdevices are allowed to be used
+       by tasks within the current cgroup.
+
+         =====         ==============================================
+         count         The total number of granted subdevices
+         list          Enumeration of the subdevices
+         =====         ==============================================
+
 GEM Buffer Ownership
 ~~~~~~~~~~~~~~~~~~~~
 
diff --git a/include/drm/drm_cgroup.h b/include/drm/drm_cgroup.h
index 2b41d4d22e33..5aac47ca536f 100644
--- a/include/drm/drm_cgroup.h
+++ b/include/drm/drm_cgroup.h
@@ -17,6 +17,9 @@ struct drmcg_props {
 
        s64                     bo_limits_total_allocated_default;
        s64                     bo_limits_peak_allocated_default;
+
+       int                     compute_capacity;
+       DECLARE_BITMAP(compute_slots, MAX_DRMCG_COMPUTE_CAPACITY);
 };
 
 void drmcg_bind(struct drm_minor (*(*acq_dm)(unsigned int minor_id)),
diff --git a/include/linux/cgroup_drm.h b/include/linux/cgroup_drm.h
index aba3b26718c0..fd02f59cabab 100644
--- a/include/linux/cgroup_drm.h
+++ b/include/linux/cgroup_drm.h
@@ -11,10 +11,14 @@
 /* limit defined per the way drm_minor_alloc operates */
 #define MAX_DRM_DEV (64 * DRM_MINOR_RENDER)
 
+#define MAX_DRMCG_COMPUTE_CAPACITY 256
+
 enum drmcg_res_type {
        DRMCG_TYPE_BO_TOTAL,
        DRMCG_TYPE_BO_PEAK,
        DRMCG_TYPE_BO_COUNT,
+       DRMCG_TYPE_COMPUTE,
+       DRMCG_TYPE_COMPUTE_EFF,
        __DRMCG_TYPE_LAST,
 };
 
@@ -32,6 +36,18 @@ struct drmcg_device_resource {
        s64                     bo_limits_peak_allocated;
 
        s64                     bo_stats_count_allocated;
+
+        /* compute_stg is used to calculate _eff before applying to _eff
+        * after considering the entire hierarchy
+        */
+       DECLARE_BITMAP(compute_stg, MAX_DRMCG_COMPUTE_CAPACITY);
+       /* user configurations */
+       s64                     compute_weight;
+       /* effective compute for the cgroup after considering
+        * relationship with other cgroup
+        */
+       s64                     compute_count_eff;
+       DECLARE_BITMAP(compute_eff, MAX_DRMCG_COMPUTE_CAPACITY);
 };
 
 /**
diff --git a/kernel/cgroup/drm.c b/kernel/cgroup/drm.c
index 62d2a9d33d0c..2eadabebdfea 100644
--- a/kernel/cgroup/drm.c
+++ b/kernel/cgroup/drm.c
@@ -9,6 +9,7 @@
 #include <linux/seq_file.h>
 #include <linux/mutex.h>
 #include <linux/kernel.h>
+#include <linux/bitmap.h>
 #include <linux/cgroup_drm.h>
 #include <drm/drm_file.h>
 #include <drm/drm_drv.h>
@@ -98,6 +99,11 @@ static inline int init_drmcg_single(struct drmcg *drmcg, 
struct drm_device *dev)
        ddr->bo_limits_peak_allocated =
                dev->drmcg_props.bo_limits_peak_allocated_default;
 
+       bitmap_copy(ddr->compute_stg, dev->drmcg_props.compute_slots,
+                       MAX_DRMCG_COMPUTE_CAPACITY);
+
+       ddr->compute_weight = CGROUP_WEIGHT_DFL;
+
        return 0;
 }
 
@@ -121,6 +127,104 @@ static inline void drmcg_update_cg_tree(struct drm_device 
*dev)
        mutex_unlock(&cgroup_mutex);
 }
 
+static void drmcg_calculate_effective_compute(struct drm_device *dev,
+               const unsigned long *free_weighted,
+               struct drmcg *parent_drmcg)
+{
+       int capacity = dev->drmcg_props.compute_capacity;
+       DECLARE_BITMAP(compute_unused, MAX_DRMCG_COMPUTE_CAPACITY);
+       DECLARE_BITMAP(compute_by_weight, MAX_DRMCG_COMPUTE_CAPACITY);
+       struct drmcg_device_resource *parent_ddr;
+       struct drmcg_device_resource *ddr;
+       int minor = dev->primary->index;
+       struct cgroup_subsys_state *pos;
+       struct drmcg *child;
+       s64 weight_sum = 0;
+       s64 unused;
+
+       parent_ddr = parent_drmcg->dev_resources[minor];
+
+       /* no static cfg, use weight for calculating the effective */
+       bitmap_copy(parent_ddr->compute_stg, free_weighted, capacity);
+
+       /* calculate compute available for dist by weight for children */
+       bitmap_copy(compute_unused, parent_ddr->compute_stg, capacity);
+       unused = bitmap_weight(compute_unused, capacity);
+       css_for_each_child(pos, &parent_drmcg->css) {
+               child = css_to_drmcg(pos);
+               ddr = child->dev_resources[minor];
+
+               /* no static allocation, participate in weight dist */
+               weight_sum += ddr->compute_weight;
+       }
+
+       css_for_each_child(pos, &parent_drmcg->css) {
+               int c;
+               int p = 0;
+               child = css_to_drmcg(pos);
+               ddr = child->dev_resources[minor];
+
+               bitmap_zero(compute_by_weight, capacity);
+               for (c = ddr->compute_weight * unused / weight_sum;
+                               c > 0; c--) {
+                       p = find_next_bit(compute_unused, capacity, p);
+                       if (p < capacity) {
+                               clear_bit(p, compute_unused);
+                               set_bit(p, compute_by_weight);
+                       }
+               }
+
+               drmcg_calculate_effective_compute(dev, compute_by_weight, 
child);
+       }
+}
+
+static void drmcg_apply_effective_compute(struct drm_device *dev)
+{
+       int capacity = dev->drmcg_props.compute_capacity;
+       int minor = dev->primary->index;
+       struct drmcg_device_resource *ddr;
+       struct cgroup_subsys_state *pos;
+       struct drmcg *drmcg;
+
+       if (root_drmcg == NULL) {
+               WARN_ON(root_drmcg == NULL);
+               return;
+       }
+
+       rcu_read_lock();
+
+       /* process the entire cgroup tree from root to simplify the algorithm */
+       drmcg_calculate_effective_compute(dev, dev->drmcg_props.compute_slots,
+                                            root_drmcg);
+
+       /* apply changes to effective only if there is a change */
+       css_for_each_descendant_pre(pos, &root_drmcg->css) {
+               drmcg = css_to_drmcg(pos);
+               ddr = drmcg->dev_resources[minor];
+
+               if (!bitmap_equal(ddr->compute_stg,
+                            ddr->compute_eff, capacity)) {
+                       bitmap_copy(ddr->compute_eff, ddr->compute_stg,
+                                capacity);
+                       ddr->compute_count_eff =
+                               bitmap_weight(ddr->compute_eff, capacity);
+               }
+       }
+       rcu_read_unlock();
+}
+
+static void drmcg_apply_effective(enum drmcg_res_type type,
+               struct drm_device *dev, struct drmcg *changed_drmcg)
+{
+       switch (type) {
+       case DRMCG_TYPE_COMPUTE:
+               drmcg_apply_effective_compute(dev);
+               break;
+       default:
+               break;
+       }
+}
+
 /**
  * drmcg_register_dev - register a DRM device for usage in drm cgroup
  * @dev: DRM device
@@ -143,7 +247,13 @@ void drmcg_register_dev(struct drm_device *dev)
        {
                dev->driver->drmcg_custom_init(dev, &dev->drmcg_props);
 
+               WARN_ON(dev->drmcg_props.compute_capacity !=
+                               bitmap_weight(dev->drmcg_props.compute_slots,
+                                       MAX_DRMCG_COMPUTE_CAPACITY));
+
                drmcg_update_cg_tree(dev);
+
+               drmcg_apply_effective(DRMCG_TYPE_COMPUTE, dev, root_drmcg);
        }
        mutex_unlock(&drmcg_mutex);
 }
@@ -297,7 +407,8 @@ static void drmcg_print_stats(struct drmcg_device_resource 
*ddr,
 }
 
 static void drmcg_print_limits(struct drmcg_device_resource *ddr,
-               struct seq_file *sf, enum drmcg_res_type type)
+               struct seq_file *sf, enum drmcg_res_type type,
+               struct drm_device *dev)
 {
        if (ddr == NULL) {
                seq_puts(sf, "\n");
@@ -311,6 +422,17 @@ static void drmcg_print_limits(struct 
drmcg_device_resource *ddr,
        case DRMCG_TYPE_BO_PEAK:
                seq_printf(sf, "%lld\n", ddr->bo_limits_peak_allocated);
                break;
+       case DRMCG_TYPE_COMPUTE:
+               seq_printf(sf, "%lld\n", ddr->compute_weight);
+               break;
+       case DRMCG_TYPE_COMPUTE_EFF:
+               seq_printf(sf, "%s=%lld %s=%*pbl\n",
+                               "count",
+                               ddr->compute_count_eff,
+                               "list",
+                               dev->drmcg_props.compute_capacity,
+                               ddr->compute_eff);
+               break;
        default:
                seq_puts(sf, "\n");
                break;
@@ -358,7 +480,7 @@ static int drmcg_seq_show_fn(int id, void *ptr, void *data)
                drmcg_print_stats(ddr, sf, type);
                break;
        case DRMCG_FTYPE_LIMIT:
-               drmcg_print_limits(ddr, sf, type);
+               drmcg_print_limits(ddr, sf, type, minor->dev);
                break;
        case DRMCG_FTYPE_DEFAULT:
                drmcg_print_default(&minor->dev->drmcg_props, sf, type);
@@ -499,9 +621,25 @@ static ssize_t drmcg_limit_write(struct kernfs_open_file 
*of, char *buf,
 
                        ddr->bo_limits_peak_allocated = val;
                        break;
+               case DRMCG_TYPE_COMPUTE:
+                       rc = drmcg_process_limit_s64_val(sattr, true,
+                               CGROUP_WEIGHT_DFL, CGROUP_WEIGHT_MAX,
+                               &val);
+
+                       if (rc || val < CGROUP_WEIGHT_MIN ||
+                                               val > CGROUP_WEIGHT_MAX) {
+                               drmcg_pr_cft_err(drmcg, rc, cft_name, minor);
+                               break;
+                       }
+
+                       ddr->compute_weight = val;
+                       break;
                default:
                        break;
                }
+
+               drmcg_apply_effective(type, dm->dev, drmcg);
+
                mutex_unlock(&dm->dev->drmcg_mutex);
 
                mutex_lock(&drmcg_mutex);
@@ -560,12 +698,44 @@ struct cftype files[] = {
                .private = DRMCG_CTF_PRIV(DRMCG_TYPE_BO_COUNT,
                                                DRMCG_FTYPE_STATS),
        },
+       {
+               .name = "compute.weight",
+               .seq_show = drmcg_seq_show,
+               .write = drmcg_limit_write,
+               .private = DRMCG_CTF_PRIV(DRMCG_TYPE_COMPUTE,
+                                               DRMCG_FTYPE_LIMIT),
+       },
+       {
+               .name = "compute.effective",
+               .seq_show = drmcg_seq_show,
+               .private = DRMCG_CTF_PRIV(DRMCG_TYPE_COMPUTE_EFF,
+                                               DRMCG_FTYPE_LIMIT),
+       },
        { }     /* terminate */
 };
 
+static int drmcg_online_fn(int id, void *ptr, void *data)
+{
+       struct drm_minor *minor = ptr;
+       struct drmcg *drmcg = data;
+
+       if (minor->type != DRM_MINOR_PRIMARY)
+               return 0;
+
+       drmcg_apply_effective(DRMCG_TYPE_COMPUTE, minor->dev, drmcg);
+
+       return 0;
+}
+
+static int drmcg_css_online(struct cgroup_subsys_state *css)
+{
+       return drm_minor_for_each(&drmcg_online_fn, css_to_drmcg(css));
+}
+
 struct cgroup_subsys gpu_cgrp_subsys = {
        .css_alloc      = drmcg_css_alloc,
        .css_free       = drmcg_css_free,
+       .css_online     = drmcg_css_online,
        .early_init     = false,
        .legacy_cftypes = files,
        .dfl_cftypes    = files,
@@ -585,6 +755,9 @@ void drmcg_device_early_init(struct drm_device *dev)
        dev->drmcg_props.bo_limits_total_allocated_default = S64_MAX;
        dev->drmcg_props.bo_limits_peak_allocated_default = S64_MAX;
 
+       dev->drmcg_props.compute_capacity = MAX_DRMCG_COMPUTE_CAPACITY;
+       bitmap_fill(dev->drmcg_props.compute_slots, MAX_DRMCG_COMPUTE_CAPACITY);
+
        drmcg_update_cg_tree(dev);
 }
 EXPORT_SYMBOL(drmcg_device_early_init);
-- 
2.25.0

_______________________________________________
dri-devel mailing list
dri-devel@lists.freedesktop.org
https://lists.freedesktop.org/mailman/listinfo/dri-devel

Reply via email to