/* * Cache the last found function. Yes, updates to this is racey, but * so is memory cache ;-)
*/ staticunsignedlong cached_function;
void ftrace_record_recursion(unsignedlong ip, unsignedlong parent_ip)
{ int index = 0; int i; unsignedlong old;
again: /* First check the last one recorded */ if (ip == cached_function) return;
i = atomic_read(&nr_records); /* nr_records is -1 when clearing records */
smp_mb__after_atomic(); if (i < 0) return;
/* * If there's two writers and this writer comes in second, * the cmpxchg() below to update the ip will fail. Then this * writer will try again. It is possible that index will now * be greater than nr_records. This is because the writer * that succeeded has not updated the nr_records yet. * This writer could keep trying again until the other writer * updates nr_records. But if the other writer takes an * interrupt, and that interrupt locks up that CPU, we do * not want this CPU to lock up due to the recursion protection, * and have a bug report showing this CPU as the cause of * locking up the computer. To not lose this record, this * writer will simply use the next position to update the * recursed_functions, and it will update the nr_records * accordingly.
*/ if (index < i)
index = i; if (index >= CONFIG_FTRACE_RECORD_RECURSION_SIZE) return;
for (i = index - 1; i >= 0; i--) { if (recursed_functions[i].ip == ip) {
cached_function = ip; return;
}
}
cached_function = ip;
/* * We only want to add a function if it hasn't been added before. * Add to the current location before incrementing the count. * If it fails to add, then increment the index (save in i) * and try again.
*/
old = cmpxchg(&recursed_functions[index].ip, 0, ip); if (old != 0) { /* Did something else already added this for us? */ if (old == ip) return; /* Try the next location (use i for the next index) */
index++; goto again;
}
recursed_functions[index].parent_ip = parent_ip;
/* * It's still possible that we could race with the clearing * CPU0 CPU1 * ---- ---- * ip = func * nr_records = -1; * recursed_functions[0] = 0; * i = -1 * if (i < 0) * nr_records = 0; * (new recursion detected) * recursed_functions[0] = func * cmpxchg(recursed_functions[0], * func, 0) * * But the worse that could happen is that we get a zero in * the recursed_functions array, and it's likely that "func" will * be recorded again.
*/
i = atomic_read(&nr_records);
smp_mb__after_atomic(); if (i < 0)
cmpxchg(&recursed_functions[index].ip, ip, 0); elseif (i <= index)
atomic_cmpxchg(&nr_records, i, index + 1);
}
EXPORT_SYMBOL_GPL(ftrace_record_recursion);
staticint recursed_function_open(struct inode *inode, struct file *file)
{ int ret = 0;
mutex_lock(&recursed_function_lock); /* If this file was opened for write, then erase contents */ if ((file->f_mode & FMODE_WRITE) && (file->f_flags & O_TRUNC)) { /* disable updating records */
atomic_set(&nr_records, -1);
smp_mb__after_atomic();
memset(recursed_functions, 0, sizeof(recursed_functions));
smp_wmb(); /* enable them again */
atomic_set(&nr_records, 0);
} if (file->f_mode & FMODE_READ)
ret = seq_open(file, &recursed_function_seq_ops);
mutex_unlock(&recursed_function_lock);
Die Informationen auf dieser Webseite wurden
nach bestem Wissen sorgfältig zusammengestellt. Es wird jedoch weder Vollständigkeit, noch Richtigkeit,
noch Qualität der bereit gestellten Informationen zugesichert.
Bemerkung:
Die farbliche Syntaxdarstellung und die Messung sind noch experimentell.