void perf_cpu_map__set_nr(struct perf_cpu_map *map, int nr_cpus)
 {
-       map->nr = nr_cpus;
+       RC_CHK_ACCESS(map)->nr = nr_cpus;
 }
 
 struct perf_cpu_map *perf_cpu_map__alloc(int nr_cpus)
 {
-       struct perf_cpu_map *cpus = malloc(sizeof(*cpus) + sizeof(struct perf_cpu) * nr_cpus);
+       RC_STRUCT(perf_cpu_map) *cpus = malloc(sizeof(*cpus) + sizeof(struct perf_cpu) * nr_cpus);
+       struct perf_cpu_map *result;
 
-       if (cpus != NULL) {
+       if (ADD_RC_CHK(result, cpus)) {
                cpus->nr = nr_cpus;
                refcount_set(&cpus->refcnt, 1);
-
        }
-       return cpus;
+       return result;
 }
 
 struct perf_cpu_map *perf_cpu_map__dummy_new(void)
        struct perf_cpu_map *cpus = perf_cpu_map__alloc(1);
 
        if (cpus)
-               cpus->map[0].cpu = -1;
+               RC_CHK_ACCESS(cpus)->map[0].cpu = -1;
 
        return cpus;
 }
        if (map) {
                WARN_ONCE(refcount_read(perf_cpu_map__refcnt(map)) != 0,
                          "cpu_map refcnt unbalanced\n");
-               free(map);
+               RC_CHK_FREE(map);
        }
 }
 
 struct perf_cpu_map *perf_cpu_map__get(struct perf_cpu_map *map)
 {
-       if (map)
+       struct perf_cpu_map *result;
+
+       if (RC_CHK_GET(result, map))
                refcount_inc(perf_cpu_map__refcnt(map));
-       return map;
+
+       return result;
 }
 
 void perf_cpu_map__put(struct perf_cpu_map *map)
 {
-       if (map && refcount_dec_and_test(perf_cpu_map__refcnt(map)))
-               cpu_map__delete(map);
+       if (map) {
+               if (refcount_dec_and_test(perf_cpu_map__refcnt(map)))
+                       cpu_map__delete(map);
+               else
+                       RC_CHK_PUT(map);
+       }
 }
 
 static struct perf_cpu_map *cpu_map__default_new(void)
                int i;
 
                for (i = 0; i < nr_cpus; ++i)
-                       cpus->map[i].cpu = i;
+                       RC_CHK_ACCESS(cpus)->map[i].cpu = i;
        }
 
        return cpus;
        int i, j;
 
        if (cpus != NULL) {
-               memcpy(cpus->map, tmp_cpus, payload_size);
-               qsort(cpus->map, nr_cpus, sizeof(struct perf_cpu), cmp_cpu);
+               memcpy(RC_CHK_ACCESS(cpus)->map, tmp_cpus, payload_size);
+               qsort(RC_CHK_ACCESS(cpus)->map, nr_cpus, sizeof(struct perf_cpu), cmp_cpu);
                /* Remove dups */
                j = 0;
                for (i = 0; i < nr_cpus; i++) {
-                       if (i == 0 || cpus->map[i].cpu != cpus->map[i - 1].cpu)
-                               cpus->map[j++].cpu = cpus->map[i].cpu;
+                       if (i == 0 || RC_CHK_ACCESS(cpus)->map[i].cpu != RC_CHK_ACCESS(cpus)->map[i - 1].cpu)
+                               RC_CHK_ACCESS(cpus)->map[j++].cpu = RC_CHK_ACCESS(cpus)->map[i].cpu;
                }
-               cpus->nr = j;
+               perf_cpu_map__set_nr(cpus, j);
                assert(j <= nr_cpus);
        }
        return cpus;
                .cpu = -1
        };
 
-       if (cpus && idx < cpus->nr)
-               return cpus->map[idx];
+       if (cpus && idx < RC_CHK_ACCESS(cpus)->nr)
+               return RC_CHK_ACCESS(cpus)->map[idx];
 
        return result;
 }
 
 int perf_cpu_map__nr(const struct perf_cpu_map *cpus)
 {
-       return cpus ? cpus->nr : 1;
+       return cpus ? RC_CHK_ACCESS(cpus)->nr : 1;
 }
 
 bool perf_cpu_map__empty(const struct perf_cpu_map *map)
 {
-       return map ? map->map[0].cpu == -1 : true;
+       return map ? RC_CHK_ACCESS(map)->map[0].cpu == -1 : true;
 }
 
 int perf_cpu_map__idx(const struct perf_cpu_map *cpus, struct perf_cpu cpu)
                return -1;
 
        low = 0;
-       high = cpus->nr;
+       high = RC_CHK_ACCESS(cpus)->nr;
        while (low < high) {
                int idx = (low + high) / 2;
-               struct perf_cpu cpu_at_idx = cpus->map[idx];
+               struct perf_cpu cpu_at_idx = RC_CHK_ACCESS(cpus)->map[idx];
 
                if (cpu_at_idx.cpu == cpu.cpu)
                        return idx;
        };
 
        // cpu_map__trim_new() qsort()s it, cpu_map__default_new() sorts it as well.
-       return map->nr > 0 ? map->map[map->nr - 1] : result;
+       return RC_CHK_ACCESS(map)->nr > 0 ? RC_CHK_ACCESS(map)->map[RC_CHK_ACCESS(map)->nr - 1] : result;
 }
 
 /** Is 'b' a subset of 'a'. */
 {
        if (a == b || !b)
                return true;
-       if (!a || b->nr > a->nr)
+       if (!a || RC_CHK_ACCESS(b)->nr > RC_CHK_ACCESS(a)->nr)
                return false;
 
-       for (int i = 0, j = 0; i < a->nr; i++) {
-               if (a->map[i].cpu > b->map[j].cpu)
+       for (int i = 0, j = 0; i < RC_CHK_ACCESS(a)->nr; i++) {
+               if (RC_CHK_ACCESS(a)->map[i].cpu > RC_CHK_ACCESS(b)->map[j].cpu)
                        return false;
-               if (a->map[i].cpu == b->map[j].cpu) {
+               if (RC_CHK_ACCESS(a)->map[i].cpu == RC_CHK_ACCESS(b)->map[j].cpu) {
                        j++;
-                       if (j == b->nr)
+                       if (j == RC_CHK_ACCESS(b)->nr)
                                return true;
                }
        }
                return perf_cpu_map__get(other);
        }
 
-       tmp_len = orig->nr + other->nr;
+       tmp_len = RC_CHK_ACCESS(orig)->nr + RC_CHK_ACCESS(other)->nr;
        tmp_cpus = malloc(tmp_len * sizeof(struct perf_cpu));
        if (!tmp_cpus)
                return NULL;
 
        /* Standard merge algorithm from wikipedia */
        i = j = k = 0;
-       while (i < orig->nr && j < other->nr) {
-               if (orig->map[i].cpu <= other->map[j].cpu) {
-                       if (orig->map[i].cpu == other->map[j].cpu)
+       while (i < RC_CHK_ACCESS(orig)->nr && j < RC_CHK_ACCESS(other)->nr) {
+               if (RC_CHK_ACCESS(orig)->map[i].cpu <= RC_CHK_ACCESS(other)->map[j].cpu) {
+                       if (RC_CHK_ACCESS(orig)->map[i].cpu == RC_CHK_ACCESS(other)->map[j].cpu)
                                j++;
-                       tmp_cpus[k++] = orig->map[i++];
+                       tmp_cpus[k++] = RC_CHK_ACCESS(orig)->map[i++];
                } else
-                       tmp_cpus[k++] = other->map[j++];
+                       tmp_cpus[k++] = RC_CHK_ACCESS(other)->map[j++];
        }
 
-       while (i < orig->nr)
-               tmp_cpus[k++] = orig->map[i++];
+       while (i < RC_CHK_ACCESS(orig)->nr)
+               tmp_cpus[k++] = RC_CHK_ACCESS(orig)->map[i++];
 
-       while (j < other->nr)
-               tmp_cpus[k++] = other->map[j++];
+       while (j < RC_CHK_ACCESS(other)->nr)
+               tmp_cpus[k++] = RC_CHK_ACCESS(other)->map[j++];
        assert(k <= tmp_len);
 
        merged = cpu_map__trim_new(k, tmp_cpus);