magazine_medium_test.c [plain text]
#include <darwintest.h>
#include "../src/magazine_medium.c"
#include "magazine_testing.h"
bool aggressive_madvise_enabled = false;
T_GLOBAL_META(T_META_RUN_CONCURRENTLY(true));
static inline void
medium_test_rack_setup(rack_t *rack)
{
test_rack_setup(rack, RACK_TYPE_MEDIUM);
}
static void
assert_block_madvise_headers(void *ptr, msize_t msize, bool dirty, bool intrusive_free_list)
{
msize_t *madv_headers = MEDIUM_MADVISE_HEADER_FOR_PTR(ptr);
msize_t index = MEDIUM_META_INDEX_FOR_PTR(ptr);
msize_t orig_msize = msize;
uintptr_t safe_start_ptr = (uintptr_t) ptr;
uintptr_t end_ptr = (uintptr_t)ptr + (msize << SHIFT_MEDIUM_QUANTUM);
uintptr_t safe_end_ptr = end_ptr;
if (intrusive_free_list && !dirty) {
safe_start_ptr = round_page_kernel((uintptr_t)ptr + sizeof(medium_inplace_free_entry_s) + sizeof(msize_t));
safe_end_ptr = trunc_page_kernel((uintptr_t) ptr + (msize << SHIFT_MEDIUM_QUANTUM) - sizeof(msize_t));
}
index = MEDIUM_META_INDEX_FOR_PTR(safe_start_ptr);
msize = MEDIUM_META_INDEX_FOR_PTR(safe_end_ptr) - index;
msize_t end_index = MEDIUM_META_INDEX_FOR_PTR(safe_end_ptr);
msize_t expected = msize | (dirty ? 0 : MEDIUM_IS_ADVISED);
T_ASSERT_EQ(madv_headers[index], expected, "Start of block is marked correctly");
if (msize > 1) {
T_ASSERT_EQ(madv_headers[end_index - 1], expected, "End of block is marked correctly");
}
for (msize_t i = 1; i < msize - 1; i++) {
T_QUIET; T_ASSERT_EQ(madv_headers[index + i], (msize_t) 0, "Middle of block is marked correctly");
}
if (intrusive_free_list) {
index = MEDIUM_META_INDEX_FOR_PTR(ptr);
if (MEDIUM_META_INDEX_FOR_PTR(safe_start_ptr) > MEDIUM_META_INDEX_FOR_PTR(ptr)) {
msize_t first_page_header = madv_headers[index];
T_ASSERT_NE(first_page_header, (msize_t)0, "free list is not marked as middle");
T_ASSERT_EQ(first_page_header & MEDIUM_IS_ADVISED, 0, "free list is marked as dirty");
}
if (MEDIUM_META_INDEX_FOR_PTR(safe_end_ptr) < MEDIUM_META_INDEX_FOR_PTR(end_ptr)) {
msize_t last_page_header = madv_headers[MEDIUM_META_INDEX_FOR_PTR(safe_end_ptr)];
T_ASSERT_NE(last_page_header, (msize_t)0, "free list is not marked as middle");
T_ASSERT_EQ(last_page_header & MEDIUM_IS_ADVISED, 0, "free list is marked as dirty");
}
}
}
static inline magazine_t *
get_magazine(struct rack_s *rack, void *ptr)
{
mag_index_t mag_index = MAGAZINE_INDEX_FOR_MEDIUM_REGION(MEDIUM_REGION_FOR_PTR(ptr));
return &(rack->magazines[mag_index]);
}
T_DECL(medium_realloc_madvise_headers, "medium realloc in place maintains madvise headers",
T_META_ENABLED(CONFIG_MEDIUM_ALLOCATOR))
{
struct rack_s rack;
medium_test_rack_setup(&rack);
void *ptr = medium_malloc_should_clear(&rack, 1, false);
T_ASSERT_NOTNULL(ptr, "allocation");
void *ptr2 = medium_malloc_should_clear(&rack, 4, false);
T_ASSERT_NOTNULL(ptr2, "allocation 2");
T_ASSERT_EQ_PTR(ptr2, (void *)((uintptr_t)ptr + MEDIUM_BYTES_FOR_MSIZE(1)), "sequential allocations");
void *extra_ptr = medium_malloc_should_clear(&rack, 1, false);
free_medium(&rack, ptr2, MEDIUM_REGION_FOR_PTR(ptr2), 0);
free_medium(&rack, extra_ptr, MEDIUM_REGION_FOR_PTR(extra_ptr), 0);
boolean_t realloced = medium_try_realloc_in_place(&rack, ptr, MEDIUM_BYTES_FOR_MSIZE(1), MEDIUM_BYTES_FOR_MSIZE(2));
T_ASSERT_TRUE(realloced, "realloced");
assert_block_madvise_headers(ptr, 2, true, false);
void *next_block = (unsigned char *)ptr + MEDIUM_BYTES_FOR_MSIZE(2);
assert_block_madvise_headers(next_block, 3, true, false);
}
T_DECL(free_end_of_region, "End of region's footer is marked dirty",
T_META_ENABLED(CONFIG_MEDIUM_ALLOCATOR))
{
struct rack_s rack;
medium_test_rack_setup(&rack);
void *oob_ptrs[MEDIUM_OOB_COUNT];
for (size_t i = 0; i < MEDIUM_OOB_COUNT * 2; i++) {
void *ptr = medium_malloc_should_clear(&rack, 1, false);
T_QUIET; T_ASSERT_NOTNULL(ptr, "oob allocation");
if (i % 2 == 0){
oob_ptrs[i / 2] = ptr;
}
}
for (size_t i = 0; i < MEDIUM_OOB_COUNT; i++){
void *ptr = oob_ptrs[i];
free_medium(&rack, ptr, MEDIUM_REGION_FOR_PTR(ptr), 0);
}
size_t num_allocated = MEDIUM_OOB_COUNT * 2;
magazine_t *mag = get_magazine(&rack, oob_ptrs[0]);
void *ptr = NULL, *last_ptr = NULL;
size_t block_size = 0, final_block_size = 0;
while (num_allocated < NUM_MEDIUM_BLOCKS) {
size_t curr_block_size = ((medium_sliding_madvise_granularity(mag)) >> SHIFT_MEDIUM_QUANTUM) - 1;
if (curr_block_size + num_allocated >= NUM_MEDIUM_BLOCKS) {
curr_block_size = NUM_MEDIUM_BLOCKS - num_allocated;
final_block_size = curr_block_size;
} else {
block_size = curr_block_size;
}
last_ptr = ptr;
ptr = medium_malloc_should_clear(&rack, curr_block_size, false);
T_QUIET; T_ASSERT_NOTNULL(ptr, "allocation under madvise window");
num_allocated += curr_block_size;
}
free_medium(&rack, last_ptr, MEDIUM_REGION_FOR_PTR(last_ptr), 0);
free_medium(&rack, ptr, MEDIUM_REGION_FOR_PTR(ptr), 0);
void *before_trailing_2 = (void *) ((uintptr_t)oob_ptrs[0] + (1UL << SHIFT_MEDIUM_QUANTUM));
free_medium(&rack, before_trailing_2, MEDIUM_REGION_FOR_PTR(before_trailing_2), 0);
assert_block_madvise_headers(last_ptr, block_size + final_block_size, false, true);
}