Browse Source
Add a test that locklessly changes and exercises page protection bits from various threads. This helps catch race conditions in the VMA handling. Acked-by: Alex Bennée <alex.bennee@linaro.org> Signed-off-by: Ilya Leoshkevich <iii@linux.ibm.com> Message-Id: <20221223120252.513319-1-iii@linux.ibm.com> Signed-off-by: Richard Henderson <richard.henderson@linaro.org>pull/229/head
committed by
Richard Henderson
4 changed files with 236 additions and 15 deletions
@ -0,0 +1,25 @@ |
|||
/*
|
|||
* No-op functions that can be safely copied. |
|||
* |
|||
* SPDX-License-Identifier: GPL-2.0-or-later |
|||
*/ |
|||
#ifndef NOP_FUNC_H |
|||
#define NOP_FUNC_H |
|||
|
|||
static const char nop_func[] = { |
|||
#if defined(__aarch64__) |
|||
0xc0, 0x03, 0x5f, 0xd6, /* ret */ |
|||
#elif defined(__alpha__) |
|||
0x01, 0x80, 0xFA, 0x6B, /* ret */ |
|||
#elif defined(__arm__) |
|||
0x1e, 0xff, 0x2f, 0xe1, /* bx lr */ |
|||
#elif defined(__riscv) |
|||
0x67, 0x80, 0x00, 0x00, /* ret */ |
|||
#elif defined(__s390__) |
|||
0x07, 0xfe, /* br %r14 */ |
|||
#elif defined(__i386__) || defined(__x86_64__) |
|||
0xc3, /* ret */ |
|||
#endif |
|||
}; |
|||
|
|||
#endif |
|||
@ -0,0 +1,207 @@ |
|||
/*
|
|||
* Test that VMA updates do not race. |
|||
* |
|||
* SPDX-License-Identifier: GPL-2.0-or-later |
|||
* |
|||
* Map a contiguous chunk of RWX memory. Split it into 8 equally sized |
|||
* regions, each of which is guaranteed to have a certain combination of |
|||
* protection bits set. |
|||
* |
|||
* Reader, writer and executor threads perform the respective operations on |
|||
* pages, which are guaranteed to have the respective protection bit set. |
|||
* Two mutator threads change the non-fixed protection bits randomly. |
|||
*/ |
|||
#include <assert.h> |
|||
#include <fcntl.h> |
|||
#include <pthread.h> |
|||
#include <stdbool.h> |
|||
#include <stdlib.h> |
|||
#include <string.h> |
|||
#include <stdio.h> |
|||
#include <sys/mman.h> |
|||
#include <unistd.h> |
|||
|
|||
#include "nop_func.h" |
|||
|
|||
#define PAGE_IDX_BITS 10 |
|||
#define PAGE_COUNT (1 << PAGE_IDX_BITS) |
|||
#define PAGE_IDX_MASK (PAGE_COUNT - 1) |
|||
#define REGION_IDX_BITS 3 |
|||
#define PAGE_IDX_R_MASK (1 << 7) |
|||
#define PAGE_IDX_W_MASK (1 << 8) |
|||
#define PAGE_IDX_X_MASK (1 << 9) |
|||
#define REGION_MASK (PAGE_IDX_R_MASK | PAGE_IDX_W_MASK | PAGE_IDX_X_MASK) |
|||
#define PAGES_PER_REGION (1 << (PAGE_IDX_BITS - REGION_IDX_BITS)) |
|||
|
|||
struct context { |
|||
int pagesize; |
|||
char *ptr; |
|||
int dev_null_fd; |
|||
volatile int mutator_count; |
|||
}; |
|||
|
|||
static void *thread_read(void *arg) |
|||
{ |
|||
struct context *ctx = arg; |
|||
ssize_t sret; |
|||
size_t i, j; |
|||
int ret; |
|||
|
|||
for (i = 0; ctx->mutator_count; i++) { |
|||
char *p; |
|||
|
|||
j = (i & PAGE_IDX_MASK) | PAGE_IDX_R_MASK; |
|||
p = &ctx->ptr[j * ctx->pagesize]; |
|||
|
|||
/* Read directly. */ |
|||
ret = memcmp(p, nop_func, sizeof(nop_func)); |
|||
if (ret != 0) { |
|||
fprintf(stderr, "fail direct read %p\n", p); |
|||
abort(); |
|||
} |
|||
|
|||
/* Read indirectly. */ |
|||
sret = write(ctx->dev_null_fd, p, 1); |
|||
if (sret != 1) { |
|||
if (sret < 0) { |
|||
fprintf(stderr, "fail indirect read %p (%m)\n", p); |
|||
} else { |
|||
fprintf(stderr, "fail indirect read %p (%zd)\n", p, sret); |
|||
} |
|||
abort(); |
|||
} |
|||
} |
|||
|
|||
return NULL; |
|||
} |
|||
|
|||
static void *thread_write(void *arg) |
|||
{ |
|||
struct context *ctx = arg; |
|||
struct timespec *ts; |
|||
size_t i, j; |
|||
int ret; |
|||
|
|||
for (i = 0; ctx->mutator_count; i++) { |
|||
j = (i & PAGE_IDX_MASK) | PAGE_IDX_W_MASK; |
|||
|
|||
/* Write directly. */ |
|||
memcpy(&ctx->ptr[j * ctx->pagesize], nop_func, sizeof(nop_func)); |
|||
|
|||
/* Write using a syscall. */ |
|||
ts = (struct timespec *)(&ctx->ptr[(j + 1) * ctx->pagesize] - |
|||
sizeof(struct timespec)); |
|||
ret = clock_gettime(CLOCK_REALTIME, ts); |
|||
if (ret != 0) { |
|||
fprintf(stderr, "fail indirect write %p (%m)\n", ts); |
|||
abort(); |
|||
} |
|||
} |
|||
|
|||
return NULL; |
|||
} |
|||
|
|||
static void *thread_execute(void *arg) |
|||
{ |
|||
struct context *ctx = arg; |
|||
size_t i, j; |
|||
|
|||
for (i = 0; ctx->mutator_count; i++) { |
|||
j = (i & PAGE_IDX_MASK) | PAGE_IDX_X_MASK; |
|||
((void(*)(void))&ctx->ptr[j * ctx->pagesize])(); |
|||
} |
|||
|
|||
return NULL; |
|||
} |
|||
|
|||
static void *thread_mutate(void *arg) |
|||
{ |
|||
size_t i, start_idx, end_idx, page_idx, tmp; |
|||
struct context *ctx = arg; |
|||
unsigned int seed; |
|||
int prot, ret; |
|||
|
|||
seed = (unsigned int)time(NULL); |
|||
for (i = 0; i < 10000; i++) { |
|||
start_idx = rand_r(&seed) & PAGE_IDX_MASK; |
|||
end_idx = rand_r(&seed) & PAGE_IDX_MASK; |
|||
if (start_idx > end_idx) { |
|||
tmp = start_idx; |
|||
start_idx = end_idx; |
|||
end_idx = tmp; |
|||
} |
|||
prot = rand_r(&seed) & (PROT_READ | PROT_WRITE | PROT_EXEC); |
|||
for (page_idx = start_idx & REGION_MASK; page_idx <= end_idx; |
|||
page_idx += PAGES_PER_REGION) { |
|||
if (page_idx & PAGE_IDX_R_MASK) { |
|||
prot |= PROT_READ; |
|||
} |
|||
if (page_idx & PAGE_IDX_W_MASK) { |
|||
/* FIXME: qemu syscalls check for both read+write. */ |
|||
prot |= PROT_WRITE | PROT_READ; |
|||
} |
|||
if (page_idx & PAGE_IDX_X_MASK) { |
|||
prot |= PROT_EXEC; |
|||
} |
|||
} |
|||
ret = mprotect(&ctx->ptr[start_idx * ctx->pagesize], |
|||
(end_idx - start_idx + 1) * ctx->pagesize, prot); |
|||
assert(ret == 0); |
|||
} |
|||
|
|||
__atomic_fetch_sub(&ctx->mutator_count, 1, __ATOMIC_SEQ_CST); |
|||
|
|||
return NULL; |
|||
} |
|||
|
|||
int main(void) |
|||
{ |
|||
pthread_t threads[5]; |
|||
struct context ctx; |
|||
size_t i; |
|||
int ret; |
|||
|
|||
/* Without a template, nothing to test. */ |
|||
if (sizeof(nop_func) == 0) { |
|||
return EXIT_SUCCESS; |
|||
} |
|||
|
|||
/* Initialize memory chunk. */ |
|||
ctx.pagesize = getpagesize(); |
|||
ctx.ptr = mmap(NULL, PAGE_COUNT * ctx.pagesize, |
|||
PROT_READ | PROT_WRITE | PROT_EXEC, |
|||
MAP_PRIVATE | MAP_ANONYMOUS, -1, 0); |
|||
assert(ctx.ptr != MAP_FAILED); |
|||
for (i = 0; i < PAGE_COUNT; i++) { |
|||
memcpy(&ctx.ptr[i * ctx.pagesize], nop_func, sizeof(nop_func)); |
|||
} |
|||
ctx.dev_null_fd = open("/dev/null", O_WRONLY); |
|||
assert(ctx.dev_null_fd >= 0); |
|||
ctx.mutator_count = 2; |
|||
|
|||
/* Start threads. */ |
|||
ret = pthread_create(&threads[0], NULL, thread_read, &ctx); |
|||
assert(ret == 0); |
|||
ret = pthread_create(&threads[1], NULL, thread_write, &ctx); |
|||
assert(ret == 0); |
|||
ret = pthread_create(&threads[2], NULL, thread_execute, &ctx); |
|||
assert(ret == 0); |
|||
for (i = 3; i <= 4; i++) { |
|||
ret = pthread_create(&threads[i], NULL, thread_mutate, &ctx); |
|||
assert(ret == 0); |
|||
} |
|||
|
|||
/* Wait for threads to stop. */ |
|||
for (i = 0; i < sizeof(threads) / sizeof(threads[0]); i++) { |
|||
ret = pthread_join(threads[i], NULL); |
|||
assert(ret == 0); |
|||
} |
|||
|
|||
/* Destroy memory chunk. */ |
|||
ret = close(ctx.dev_null_fd); |
|||
assert(ret == 0); |
|||
ret = munmap(ctx.ptr, PAGE_COUNT * ctx.pagesize); |
|||
assert(ret == 0); |
|||
|
|||
return EXIT_SUCCESS; |
|||
} |
|||
Loading…
Reference in new issue