unsigned long i;
ret = mtree_test_store_range(mt, start, end, ptr);
- printk("ret %d expected %d\n", ret, expected);
MT_BUG_ON(mt, ret != expected);
if (ret)
xa_mk_value(range[i] >> 12), 0);
}
- printk("Starting reverse range allocation tests.\n");
for (i = 0; i < req_range_cnt; i += 5) {
- mt_dump(mt);
- printk("%s %d: alloc %ld in range %ld - %ld\n",
- __func__, i,
- req_range[i+2] >> 12,
- req_range[i] >> 12,
- req_range[i+1] >> 12);
- if (req_range[i+4])
- printk("\t Should fail\n");
- else
- printk ("\tShould be starting at %ld\n",
- req_range[i+3] >> 12);
check_mtree_alloc_rrange(mt,
req_range[i] >> 12, // start
req_range[i+1] >> 12, // end
req_range[i+3] >> 12, // expected address
req_range[i+4], // expected return
xa_mk_value(req_range[i] >> 12)); // pointer
- mt_dump(mt);
- printk("Done\n\n");
}
mtree_destroy(mt);
MT_BUG_ON(mt, !mtree_empty(mt));
for (i = 0; i < range_cnt; i += 2) {
- mt_dump(mt);
- printk("Starting check insert %d\n", i);
check_insert_range(mt, range[i] >> 12, (range[i + 1] >> 12) - 1,
xa_mk_value(range[i] >> 12), 0);
}
for (i = 0; i < req_range_cnt; i += 5) {
- mt_dump(mt);
- printk("%s %d: alloc %ld in range %ld - %ld\n",
- __func__, i,
- req_range[i+2] >> 12,
- req_range[i] >> 12,
- req_range[i+1] >> 12);
- if (req_range[i+4])
- printk("\t Should fail\n");
- else
- printk ("\tShould be starting at %ld\n",
- req_range[i+3] >> 12);
check_mtree_alloc_range(mt,
req_range[i] >> 12, // start
req_range[i+1] >> 12, // end
req_range[i+3] >> 12, // expected address
req_range[i+4], // expected return
xa_mk_value(req_range[i] >> 12)); // pointer
- mt_dump(mt);
- printk("Done\n\n");
}
mtree_destroy(mt);
45, 168,
};
- printk("Check ranges\n");
MT_BUG_ON(mt, !mtree_empty(mt));
check_insert_range(mt, r[0], r[1], xa_mk_value(r[0]), 0);
check_insert_range(mt, r[2], r[3], xa_mk_value(r[2]), 0);
MT_BUG_ON(mt, !mtree_empty(mt));
check_seq(mt, 30, false);
- mt_dump(mt);
mtree_destroy(mt);
}