Loading...
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 | // SPDX-License-Identifier: GPL-2.0 #include <linux/seq_file.h> #include <linux/kallsyms.h> #include <linux/module.h> #include <linux/ftrace.h> #include <linux/fs.h> #include "trace_output.h" struct recursed_functions { unsigned long ip; unsigned long parent_ip; }; static struct recursed_functions recursed_functions[CONFIG_FTRACE_RECORD_RECURSION_SIZE]; static atomic_t nr_records; /* * Cache the last found function. Yes, updates to this is racey, but * so is memory cache ;-) */ static unsigned long cached_function; void ftrace_record_recursion(unsigned long ip, unsigned long parent_ip) { int index = 0; int i; unsigned long old; again: /* First check the last one recorded */ if (ip == cached_function) return; i = atomic_read(&nr_records); /* nr_records is -1 when clearing records */ smp_mb__after_atomic(); if (i < 0) return; /* * If there's two writers and this writer comes in second, * the cmpxchg() below to update the ip will fail. Then this * writer will try again. It is possible that index will now * be greater than nr_records. This is because the writer * that succeeded has not updated the nr_records yet. * This writer could keep trying again until the other writer * updates nr_records. But if the other writer takes an * interrupt, and that interrupt locks up that CPU, we do * not want this CPU to lock up due to the recursion protection, * and have a bug report showing this CPU as the cause of * locking up the computer. To not lose this record, this * writer will simply use the next position to update the * recursed_functions, and it will update the nr_records * accordingly. */ if (index < i) index = i; if (index >= CONFIG_FTRACE_RECORD_RECURSION_SIZE) return; for (i = index - 1; i >= 0; i--) { if (recursed_functions[i].ip == ip) { cached_function = ip; return; } } cached_function = ip; /* * We only want to add a function if it hasn't been added before. * Add to the current location before incrementing the count. * If it fails to add, then increment the index (save in i) * and try again. */ old = cmpxchg(&recursed_functions[index].ip, 0, ip); if (old != 0) { /* Did something else already added this for us? */ if (old == ip) return; /* Try the next location (use i for the next index) */ index++; goto again; } recursed_functions[index].parent_ip = parent_ip; /* * It's still possible that we could race with the clearing * CPU0 CPU1 * ---- ---- * ip = func * nr_records = -1; * recursed_functions[0] = 0; * i = -1 * if (i < 0) * nr_records = 0; * (new recursion detected) * recursed_functions[0] = func * cmpxchg(recursed_functions[0], * func, 0) * * But the worse that could happen is that we get a zero in * the recursed_functions array, and it's likely that "func" will * be recorded again. */ i = atomic_read(&nr_records); smp_mb__after_atomic(); if (i < 0) cmpxchg(&recursed_functions[index].ip, ip, 0); else if (i <= index) atomic_cmpxchg(&nr_records, i, index + 1); } EXPORT_SYMBOL_GPL(ftrace_record_recursion); static DEFINE_MUTEX(recursed_function_lock); static struct trace_seq *tseq; static void *recursed_function_seq_start(struct seq_file *m, loff_t *pos) { void *ret = NULL; int index; mutex_lock(&recursed_function_lock); index = atomic_read(&nr_records); if (*pos < index) { ret = &recursed_functions[*pos]; } tseq = kzalloc(sizeof(*tseq), GFP_KERNEL); if (!tseq) return ERR_PTR(-ENOMEM); trace_seq_init(tseq); return ret; } static void *recursed_function_seq_next(struct seq_file *m, void *v, loff_t *pos) { int index; int p; index = atomic_read(&nr_records); p = ++(*pos); return p < index ? &recursed_functions[p] : NULL; } static void recursed_function_seq_stop(struct seq_file *m, void *v) { kfree(tseq); mutex_unlock(&recursed_function_lock); } static int recursed_function_seq_show(struct seq_file *m, void *v) { struct recursed_functions *record = v; int ret = 0; if (record) { trace_seq_print_sym(tseq, record->parent_ip, true); trace_seq_puts(tseq, ":\t"); trace_seq_print_sym(tseq, record->ip, true); trace_seq_putc(tseq, '\n'); ret = trace_print_seq(m, tseq); } return ret; } static const struct seq_operations recursed_function_seq_ops = { .start = recursed_function_seq_start, .next = recursed_function_seq_next, .stop = recursed_function_seq_stop, .show = recursed_function_seq_show }; static int recursed_function_open(struct inode *inode, struct file *file) { int ret = 0; mutex_lock(&recursed_function_lock); /* If this file was opened for write, then erase contents */ if ((file->f_mode & FMODE_WRITE) && (file->f_flags & O_TRUNC)) { /* disable updating records */ atomic_set(&nr_records, -1); smp_mb__after_atomic(); memset(recursed_functions, 0, sizeof(recursed_functions)); smp_wmb(); /* enable them again */ atomic_set(&nr_records, 0); } if (file->f_mode & FMODE_READ) ret = seq_open(file, &recursed_function_seq_ops); mutex_unlock(&recursed_function_lock); return ret; } static ssize_t recursed_function_write(struct file *file, const char __user *buffer, size_t count, loff_t *ppos) { return count; } static int recursed_function_release(struct inode *inode, struct file *file) { if (file->f_mode & FMODE_READ) seq_release(inode, file); return 0; } static const struct file_operations recursed_functions_fops = { .open = recursed_function_open, .write = recursed_function_write, .read = seq_read, .llseek = seq_lseek, .release = recursed_function_release, }; __init static int create_recursed_functions(void) { struct dentry *dentry; dentry = trace_create_file("recursed_functions", TRACE_MODE_WRITE, NULL, NULL, &recursed_functions_fops); if (!dentry) pr_warn("WARNING: Failed to create recursed_functions\n"); return 0; } fs_initcall(create_recursed_functions); |