int rdtgroup_mondata_show(struct seq_file *m, void *arg)
 {
        struct kernfs_open_file *of = m->private;
-       struct rdt_hw_resource *hw_res;
        u32 resid, evtid, domid;
        struct rdtgroup *rdtgrp;
        struct rdt_resource *r;
        domid = md.u.domid;
        evtid = md.u.evtid;
 
-       hw_res = &rdt_resources_all[resid];
-       r = &hw_res->r_resctrl;
+       r = &rdt_resources_all[resid].r_resctrl;
        d = rdt_find_domain(r, domid, NULL);
        if (IS_ERR_OR_NULL(d)) {
                ret = -ENOENT;
        else if (rr.err == -EINVAL)
                seq_puts(m, "Unavailable\n");
        else
-               seq_printf(m, "%llu\n", rr.val * hw_res->mon_scale);
+               seq_printf(m, "%llu\n", rr.val);
 
 out:
        rdtgroup_kn_unlock(of->kn);
 
  */
 
 #include <linux/module.h>
+#include <linux/sizes.h>
 #include <linux/slab.h>
 
 #include <asm/cpu_device_id.h>
        struct rdt_hw_resource *hw_res = resctrl_to_arch_res(r);
        struct rdt_hw_domain *hw_dom = resctrl_to_arch_dom(d);
        struct arch_mbm_state *am;
-       u64 msr_val;
+       u64 msr_val, chunks;
 
        if (!cpumask_test_cpu(smp_processor_id(), &d->cpu_mask))
                return -EINVAL;
        if (am) {
                am->chunks += mbm_overflow_count(am->prev_msr, msr_val,
                                                 hw_res->mbm_width);
-               *val = get_corrected_mbm_count(rmid, am->chunks);
+               chunks = get_corrected_mbm_count(rmid, am->chunks);
                am->prev_msr = msr_val;
        } else {
-               *val = msr_val;
+               chunks = msr_val;
        }
 
+       *val = chunks * hw_res->mon_scale;
+
        return 0;
 }
 
 void __check_limbo(struct rdt_domain *d, bool force_free)
 {
        struct rdt_resource *r = &rdt_resources_all[RDT_RESOURCE_L3].r_resctrl;
-       struct rdt_hw_resource *hw_res = resctrl_to_arch_res(r);
        struct rmid_entry *entry;
        u32 crmid = 1, nrmid;
        bool rmid_dirty;
                                           QOS_L3_OCCUP_EVENT_ID, &val)) {
                        rmid_dirty = true;
                } else {
-                       val *= hw_res->mon_scale;
                        rmid_dirty = (val >= resctrl_rmid_realloc_threshold);
                }
 
 static void add_rmid_to_limbo(struct rmid_entry *entry)
 {
        struct rdt_resource *r = &rdt_resources_all[RDT_RESOURCE_L3].r_resctrl;
-       struct rdt_hw_resource *hw_res = resctrl_to_arch_res(r);
        struct rdt_domain *d;
        int cpu, err;
        u64 val = 0;
                        err = resctrl_arch_rmid_read(r, d, entry->rmid,
                                                     QOS_L3_OCCUP_EVENT_ID,
                                                     &val);
-                       val *= hw_res->mon_scale;
                        if (err || val <= resctrl_rmid_realloc_threshold)
                                continue;
                }
  */
 static void mbm_bw_count(u32 rmid, struct rmid_read *rr)
 {
-       struct rdt_hw_resource *hw_res = resctrl_to_arch_res(rr->r);
        struct mbm_state *m = &rr->d->mbm_local[rmid];
-       u64 cur_bw, chunks, cur_chunks;
+       u64 cur_bw, bytes, cur_bytes;
 
-       cur_chunks = rr->val;
-       chunks = cur_chunks - m->prev_bw_chunks;
-       m->prev_bw_chunks = cur_chunks;
+       cur_bytes = rr->val;
+       bytes = cur_bytes - m->prev_bw_bytes;
+       m->prev_bw_bytes = cur_bytes;
 
-       cur_bw = (chunks * hw_res->mon_scale) >> 20;
+       cur_bw = bytes / SZ_1M;
 
        if (m->delta_comp)
                m->delta_bw = abs(cur_bw - m->prev_bw);