Loading...
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 | // SPDX-License-Identifier: GPL-2.0 // Copyright (C) 2020 ARM Limited #define _GNU_SOURCE #include <errno.h> #include <fcntl.h> #include <signal.h> #include <stdlib.h> #include <stdio.h> #include <string.h> #include <ucontext.h> #include <unistd.h> #include <sys/mman.h> #include "kselftest.h" #include "mte_common_util.h" #include "mte_def.h" static size_t page_sz; static int check_usermem_access_fault(int mem_type, int mode, int mapping) { int fd, i, err; char val = 'A'; size_t len, read_len; void *ptr, *ptr_next; err = KSFT_FAIL; len = 2 * page_sz; mte_switch_mode(mode, MTE_ALLOW_NON_ZERO_TAG); fd = create_temp_file(); if (fd == -1) return KSFT_FAIL; for (i = 0; i < len; i++) write(fd, &val, sizeof(val)); lseek(fd, 0, 0); ptr = mte_allocate_memory(len, mem_type, mapping, true); if (check_allocated_memory(ptr, len, mem_type, true) != KSFT_PASS) { close(fd); return KSFT_FAIL; } mte_initialize_current_context(mode, (uintptr_t)ptr, len); /* Copy from file into buffer with valid tag */ read_len = read(fd, ptr, len); mte_wait_after_trig(); if (cur_mte_cxt.fault_valid || read_len < len) goto usermem_acc_err; /* Verify same pattern is read */ for (i = 0; i < len; i++) if (*(char *)(ptr + i) != val) break; if (i < len) goto usermem_acc_err; /* Tag the next half of memory with different value */ ptr_next = (void *)((unsigned long)ptr + page_sz); ptr_next = mte_insert_new_tag(ptr_next); mte_set_tag_address_range(ptr_next, page_sz); lseek(fd, 0, 0); /* Copy from file into buffer with invalid tag */ read_len = read(fd, ptr, len); mte_wait_after_trig(); /* * Accessing user memory in kernel with invalid tag should fail in sync * mode without fault but may not fail in async mode as per the * implemented MTE userspace support in Arm64 kernel. */ if (mode == MTE_SYNC_ERR && !cur_mte_cxt.fault_valid && read_len < len) { err = KSFT_PASS; } else if (mode == MTE_ASYNC_ERR && !cur_mte_cxt.fault_valid && read_len == len) { err = KSFT_PASS; } usermem_acc_err: mte_free_memory((void *)ptr, len, mem_type, true); close(fd); return err; } int main(int argc, char *argv[]) { int err; page_sz = getpagesize(); if (!page_sz) { ksft_print_msg("ERR: Unable to get page size\n"); return KSFT_FAIL; } err = mte_default_setup(); if (err) return err; /* Register signal handlers */ mte_register_signal(SIGSEGV, mte_default_handler); /* Set test plan */ ksft_set_plan(4); evaluate_test(check_usermem_access_fault(USE_MMAP, MTE_SYNC_ERR, MAP_PRIVATE), "Check memory access from kernel in sync mode, private mapping and mmap memory\n"); evaluate_test(check_usermem_access_fault(USE_MMAP, MTE_SYNC_ERR, MAP_SHARED), "Check memory access from kernel in sync mode, shared mapping and mmap memory\n"); evaluate_test(check_usermem_access_fault(USE_MMAP, MTE_ASYNC_ERR, MAP_PRIVATE), "Check memory access from kernel in async mode, private mapping and mmap memory\n"); evaluate_test(check_usermem_access_fault(USE_MMAP, MTE_ASYNC_ERR, MAP_SHARED), "Check memory access from kernel in async mode, shared mapping and mmap memory\n"); mte_restore_setup(); ksft_print_cnts(); return ksft_get_fail_cnt() == 0 ? KSFT_PASS : KSFT_FAIL; } |