__u32 duration = 0, retval;
        char buf[128];
        int err;
+       int page_size = getpagesize();
 
-       skel = map_ptr_kern__open_and_load();
-       if (CHECK(!skel, "skel_open_load", "open_load failed\n"))
+       skel = map_ptr_kern__open();
+       if (!ASSERT_OK_PTR(skel, "skel_open"))
                return;
 
+       err = bpf_map__set_max_entries(skel->maps.m_ringbuf, page_size);
+       if (!ASSERT_OK(err, "bpf_map__set_max_entries"))
+               goto cleanup;
+
+       err = map_ptr_kern__load(skel);
+       if (!ASSERT_OK(err, "skel_load"))
+               goto cleanup;
+
+       skel->bss->page_size = page_size;
+
        err = bpf_prog_test_run(bpf_program__fd(skel->progs.cg_skb), 1, &pkt_v4,
                                sizeof(pkt_v4), buf, NULL, &retval, NULL);
 
 
 
 enum bpf_map_type g_map_type = BPF_MAP_TYPE_UNSPEC;
 __u32 g_line = 0;
+int page_size = 0; /* userspace should set it */
 
 #define VERIFY_TYPE(type, func) ({     \
        g_map_type = type;              \
 
 struct {
        __uint(type, BPF_MAP_TYPE_RINGBUF);
-       __uint(max_entries, 1 << 12);
 } m_ringbuf SEC(".maps");
 
 static inline int check_ringbuf(void)
        struct bpf_ringbuf_map *ringbuf = (struct bpf_ringbuf_map *)&m_ringbuf;
        struct bpf_map *map = (struct bpf_map *)&m_ringbuf;
 
-       VERIFY(check(&ringbuf->map, map, 0, 0, 1 << 12));
+       VERIFY(check(&ringbuf->map, map, 0, 0, page_size));
 
        return 1;
 }