Currently we store the per package "total bytes" for each rdtgroup for
Memory bandwidth management which exposed via
"/sys/fs/resctrl/<rdtgrpx>/mon_data/mon_L3_00/mbm_local_bytes".

The above user interface remains while we also add support to measure
the per package b/w in Megabytes and the delta b/w when the b/w MSR
values change. We do this by taking the time stamp every time a the
counter is read and then keeping a history of b/w. This will be used to
support internal queries for the bandwidth in Megabytes.

Signed-off-by: Vikas Shivappa <vikas.shiva...@linux.intel.com>
---
 arch/x86/kernel/cpu/intel_rdt.c         |  1 -
 arch/x86/kernel/cpu/intel_rdt.h         | 24 ++++++++++++++++++++--
 arch/x86/kernel/cpu/intel_rdt_monitor.c | 36 +++++++++++++++++++++++++++------
 3 files changed, 52 insertions(+), 9 deletions(-)

diff --git a/arch/x86/kernel/cpu/intel_rdt.c b/arch/x86/kernel/cpu/intel_rdt.c
index 8a12d26..78beb64 100644
--- a/arch/x86/kernel/cpu/intel_rdt.c
+++ b/arch/x86/kernel/cpu/intel_rdt.c
@@ -33,7 +33,6 @@
 #include <asm/intel_rdt_sched.h>
 #include "intel_rdt.h"
 
-#define MAX_MBA_BW     100u
 #define MBA_IS_LINEAR  0x4
 #define MBA_BW_MAX_MB  U32_MAX
 
diff --git a/arch/x86/kernel/cpu/intel_rdt.h b/arch/x86/kernel/cpu/intel_rdt.h
index 68c7da0..b74619d 100644
--- a/arch/x86/kernel/cpu/intel_rdt.h
+++ b/arch/x86/kernel/cpu/intel_rdt.h
@@ -28,6 +28,18 @@
 
 #define MBM_CNTR_WIDTH                 24
 #define MBM_OVERFLOW_INTERVAL          1000
+#define MAX_MBA_BW                     100u
+
+/*
+ * This measures a tolerable delta value in MegaBytes between
+ * the expected bandwidth and the actual bandwidth.
+ * This is done so that we dont keep flipping the control
+ * bandwidth to more than and less than the expected bandwidth.
+ *
+ * However note that this is only initial threshold value and
+ * it is adjusted dynamically package wise for each rdtgrp
+ */
+#define MBA_BW_MB_THRSHL               1024
 
 #define RMID_VAL_ERROR                 BIT_ULL(63)
 #define RMID_VAL_UNAVAIL               BIT_ULL(62)
@@ -180,10 +192,18 @@ struct rftype {
  * struct mbm_state - status for each MBM counter in each domain
  * @chunks:    Total data moved (multiply by rdt_group.mon_scale to get bytes)
  * @prev_msr   Value of IA32_QM_CTR for this RMID last time we read it
+ * @prev_read_time:The last time counter was read
+ * @prev_bw:   The most recent bandwidth in Megabytes
+ * @delta_bw:  Difference between the current b/w and previous b/w
+ * @threshl_calib:     Indicates when to calculate the delta_bw
  */
 struct mbm_state {
-       u64     chunks;
-       u64     prev_msr;
+       u64                     chunks;
+       u64                     prev_msr;
+       unsigned long           prev_read_time;
+       u64                     prev_bw;
+       u64                     delta_bw;
+       bool                    thrshl_calib;
 };
 
 /**
diff --git a/arch/x86/kernel/cpu/intel_rdt_monitor.c 
b/arch/x86/kernel/cpu/intel_rdt_monitor.c
index 681450e..509f338 100644
--- a/arch/x86/kernel/cpu/intel_rdt_monitor.c
+++ b/arch/x86/kernel/cpu/intel_rdt_monitor.c
@@ -225,9 +225,11 @@ void free_rmid(u32 rmid)
                list_add_tail(&entry->list, &rmid_free_lru);
 }
 
-static int __mon_event_count(u32 rmid, struct rmid_read *rr)
+static int __mon_event_count(u32 rmid, struct rmid_read *rr, struct mbm_state 
**md)
 {
-       u64 chunks, shift, tval;
+       struct rdt_resource *r = &rdt_resources_all[RDT_RESOURCE_L3];
+       u64 chunks, shift, tval, cur_bw = 0;
+       unsigned long delta_time, now;
        struct mbm_state *m;
 
        tval = __rmid_read(rmid, rr->evtid);
@@ -256,6 +258,9 @@ static int __mon_event_count(u32 rmid, struct rmid_read *rr)
        if (rr->first) {
                m->prev_msr = tval;
                m->chunks = 0;
+               m->prev_read_time = jiffies;
+               m->prev_bw = 0;
+               m->delta_bw = MBA_BW_MB_THRSHL;
                return 0;
        }
 
@@ -266,6 +271,24 @@ static int __mon_event_count(u32 rmid, struct rmid_read 
*rr)
        m->prev_msr = tval;
 
        rr->val += m->chunks;
+
+       if(!md)
+               goto out;
+
+       now = jiffies;
+       delta_time = jiffies_to_usecs(now - m->prev_read_time);
+       if (delta_time)
+               cur_bw = (chunks * r->mon_scale) / delta_time;
+
+       if (m->thrshl_calib)
+               m->delta_bw = abs(cur_bw - m->prev_bw);
+       m->thrshl_calib = false;
+       m->prev_bw = cur_bw;
+       m->prev_read_time = now;
+
+       *md = m;
+out:
+
        return 0;
 }
 
@@ -281,7 +304,7 @@ void mon_event_count(void *info)
 
        rdtgrp = rr->rgrp;
 
-       if (__mon_event_count(rdtgrp->mon.rmid, rr))
+       if (__mon_event_count(rdtgrp->mon.rmid, rr, NULL))
                return;
 
        /*
@@ -291,7 +314,7 @@ void mon_event_count(void *info)
 
        if (rdtgrp->type == RDTCTRL_GROUP) {
                list_for_each_entry(entry, head, mon.crdtgrp_list) {
-                       if (__mon_event_count(entry->mon.rmid, rr))
+                       if (__mon_event_count(entry->mon.rmid, rr, NULL))
                                return;
                }
        }
@@ -299,6 +322,7 @@ void mon_event_count(void *info)
 
 static void mbm_update(struct rdt_domain *d, int rmid)
 {
+
        struct rmid_read rr;
 
        rr.first = false;
@@ -310,11 +334,11 @@ static void mbm_update(struct rdt_domain *d, int rmid)
         */
        if (is_mbm_total_enabled()) {
                rr.evtid = QOS_L3_MBM_TOTAL_EVENT_ID;
-               __mon_event_count(rmid, &rr);
+               __mon_event_count(rmid, &rr, NULL);
        }
        if (is_mbm_local_enabled()) {
                rr.evtid = QOS_L3_MBM_LOCAL_EVENT_ID;
-               __mon_event_count(rmid, &rr);
+               __mon_event_count(rmid, &rr, NULL);
        }
 }
 
-- 
1.9.1

Reply via email to