/* Number of 64-bit words written per one comparison: */ #define KCOV_WORDS_PER_CMP 4
/* * kcov descriptor (one per opened debugfs file). * State transitions of the descriptor: * - initial state after open() * - then there must be a single ioctl(KCOV_INIT_TRACE) call * - then, mmap() call (several calls are allowed but not useful) * - then, ioctl(KCOV_ENABLE, arg), where arg is * KCOV_TRACE_PC - to trace only the PCs * or * KCOV_TRACE_CMP - to trace only the comparison operands * - then, ioctl(KCOV_DISABLE) to disable the task. * Enabling/disabling ioctls can be repeated (only one task a time allowed).
*/ struct kcov { /* * Reference counter. We keep one for: * - opened file descriptor * - task with enabled coverage (we can't unwire it from another task) * - each code section for remote coverage collection
*/
refcount_t refcount; /* The lock protects mode, size, area and t. */
spinlock_t lock; enum kcov_mode mode; /* Size of arena (in long's). */ unsignedint size; /* Coverage buffer shared with user space. */ void *area; /* Task for which we collect coverage, or NULL. */ struct task_struct *t; /* Collecting coverage from remote (background) threads. */ bool remote; /* Size of remote area (in long's). */ unsignedint remote_size; /* * Sequence is incremented each time kcov is reenabled, used by * kcov_remote_stop(), see the comment there.
*/ int sequence;
};
/* Must be called with kcov_remote_lock locked. */ staticstruct kcov_remote *kcov_remote_add(struct kcov *kcov, u64 handle)
{ struct kcov_remote *remote;
/* Must be called with kcov_remote_lock locked. */ staticstruct kcov_remote_area *kcov_remote_area_get(unsignedint size)
{ struct kcov_remote_area *area; struct list_head *pos;
/* Must be called with kcov_remote_lock locked. */ staticvoid kcov_remote_area_put(struct kcov_remote_area *area, unsignedint size)
{
INIT_LIST_HEAD(&area->list);
area->size = size;
list_add(&area->list, &kcov_remote_areas); /* * KMSAN doesn't instrument this file, so it may not know area->list * is initialized. Unpoison it explicitly to avoid reports in * kcov_remote_area_get().
*/
kmsan_unpoison_memory(&area->list, sizeof(area->list));
}
/* * Unlike in_serving_softirq(), this function returns false when called during * a hardirq or an NMI that happened in the softirq context.
*/ static __always_inline bool in_softirq_really(void)
{ return in_serving_softirq() && !in_hardirq() && !in_nmi();
}
/* * We are interested in code coverage as a function of a syscall inputs, * so we ignore code executed in interrupts, unless we are in a remote * coverage collection section in a softirq.
*/ if (!in_task() && !(in_softirq_really() && t->kcov_softirq)) returnfalse;
mode = READ_ONCE(t->kcov_mode); /* * There is some code that runs in interrupts but for which * in_interrupt() returns false (e.g. preempt_schedule_irq()). * READ_ONCE()/barrier() effectively provides load-acquire wrt * interrupts, there are paired barrier()/WRITE_ONCE() in * kcov_start().
*/
barrier(); return mode == needed_mode;
}
/* * Entry point from instrumented code. * This is called once per basic-block/edge.
*/ void notrace __sanitizer_cov_trace_pc(void)
{ struct task_struct *t; unsignedlong *area; unsignedlong ip = canonicalize_ip(_RET_IP_); unsignedlong pos;
t = current; if (!check_kcov_mode(KCOV_MODE_TRACE_PC, t)) return;
area = t->kcov_area; /* The first 64-bit word is the number of subsequent PCs. */
pos = READ_ONCE(area[0]) + 1; if (likely(pos < t->kcov_size)) { /* Previously we write pc before updating pos. However, some * early interrupt code could bypass check_kcov_mode() check * and invoke __sanitizer_cov_trace_pc(). If such interrupt is * raised between writing pc and updating pos, the pc could be * overitten by the recursive __sanitizer_cov_trace_pc(). * Update pos before writing pc to avoid such interleaving.
*/
WRITE_ONCE(area[0], pos);
barrier();
area[pos] = ip;
}
}
EXPORT_SYMBOL(__sanitizer_cov_trace_pc);
t = current; if (!check_kcov_mode(KCOV_MODE_TRACE_CMP, t)) return;
ip = canonicalize_ip(ip);
/* * We write all comparison arguments and types as u64. * The buffer was allocated for t->kcov_size unsigned longs.
*/
area = (u64 *)t->kcov_area;
max_pos = t->kcov_size * sizeof(unsignedlong);
count = READ_ONCE(area[0]);
/* Every record is KCOV_WORDS_PER_CMP 64-bit words. */
start_index = 1 + count * KCOV_WORDS_PER_CMP;
end_pos = (start_index + KCOV_WORDS_PER_CMP) * sizeof(u64); if (likely(end_pos <= max_pos)) { /* See comment in __sanitizer_cov_trace_pc(). */
WRITE_ONCE(area[0], count + 1);
barrier();
area[start_index] = type;
area[start_index + 1] = arg1;
area[start_index + 2] = arg2;
area[start_index + 3] = ip;
}
}
switch (size) { case 8:
type |= KCOV_CMP_SIZE(0); break; case 16:
type |= KCOV_CMP_SIZE(1); break; case 32:
type |= KCOV_CMP_SIZE(2); break; case 64:
type |= KCOV_CMP_SIZE(3); break; default: return;
} for (i = 0; i < count; i++)
write_comp_data(type, cases[i + 2], val, _RET_IP_);
}
EXPORT_SYMBOL(__sanitizer_cov_trace_switch); #endif/* ifdef CONFIG_KCOV_ENABLE_COMPARISONS */
staticvoid kcov_start(struct task_struct *t, struct kcov *kcov, unsignedint size, void *area, enum kcov_mode mode, int sequence)
{
kcov_debug("t = %px, size = %u, area = %px\n", t, size, area);
t->kcov = kcov; /* Cache in task struct for performance. */
t->kcov_size = size;
t->kcov_area = area;
t->kcov_sequence = sequence; /* See comment in check_kcov_mode(). */
barrier();
WRITE_ONCE(t->kcov_mode, mode);
}
spin_lock_irqsave(&kcov->lock, flags);
kcov_debug("t = %px, kcov->t = %px\n", t, kcov->t); /* * For KCOV_ENABLE devices we want to make sure that t->kcov->t == t, * which comes down to: * WARN_ON(!kcov->remote && kcov->t != t); * * For KCOV_REMOTE_ENABLE devices, the exiting task is either: * * 1. A remote task between kcov_remote_start() and kcov_remote_stop(). * In this case we should print a warning right away, since a task * shouldn't be exiting when it's in a kcov coverage collection * section. Here t points to the task that is collecting remote * coverage, and t->kcov->t points to the thread that created the * kcov device. Which means that to detect this case we need to * check that t != t->kcov->t, and this gives us the following: * WARN_ON(kcov->remote && kcov->t != t); * * 2. The task that created kcov exiting without calling KCOV_DISABLE, * and then again we make sure that t->kcov->t == t: * WARN_ON(kcov->remote && kcov->t != t); * * By combining all three checks into one we get:
*/ if (WARN_ON(kcov->t != t)) {
spin_unlock_irqrestore(&kcov->lock, flags); return;
} /* Just to not leave dangling references behind. */
kcov_disable(t, kcov);
spin_unlock_irqrestore(&kcov->lock, flags);
kcov_put(kcov);
}
/* * Fault in a lazily-faulted vmalloc area before it can be used by * __sanitizer_cov_trace_pc(), to avoid recursion issues if any code on the * vmalloc fault handling path is instrumented.
*/ staticvoid kcov_fault_in_area(struct kcov *kcov)
{ unsignedlong stride = PAGE_SIZE / sizeof(unsignedlong); unsignedlong *area = kcov->area; unsignedlong offset;
switch (cmd) { case KCOV_ENABLE: /* * Enable coverage for the current task. * At this point user must have been enabled trace mode, * and mmapped the file. Coverage collection is disabled only * at task exit or voluntary by KCOV_DISABLE. After that it can * be enabled for another task.
*/ if (kcov->mode != KCOV_MODE_INIT || !kcov->area) return -EINVAL;
t = current; if (kcov->t != NULL || t->kcov != NULL) return -EBUSY;
mode = kcov_get_mode(arg); if (mode < 0) return mode;
kcov_fault_in_area(kcov);
kcov->mode = mode;
kcov_start(t, kcov, kcov->size, kcov->area, kcov->mode,
kcov->sequence);
kcov->t = t; /* Put either in kcov_task_exit() or in KCOV_DISABLE. */
kcov_get(kcov); return 0; case KCOV_DISABLE: /* Disable coverage for the current task. */
unused = arg; if (unused != 0 || current->kcov != kcov) return -EINVAL;
t = current; if (WARN_ON(kcov->t != t)) return -EINVAL;
kcov_disable(t, kcov);
kcov_put(kcov); return 0; case KCOV_REMOTE_ENABLE: if (kcov->mode != KCOV_MODE_INIT || !kcov->area) return -EINVAL;
t = current; if (kcov->t != NULL || t->kcov != NULL) return -EBUSY;
remote_arg = (struct kcov_remote_arg *)arg;
mode = kcov_get_mode(remote_arg->trace_mode); if (mode < 0) return mode; if ((unsignedlong)remote_arg->area_size >
LONG_MAX / sizeof(unsignedlong)) return -EINVAL;
kcov->mode = mode;
t->kcov = kcov;
t->kcov_mode = KCOV_MODE_REMOTE;
kcov->t = t;
kcov->remote = true;
kcov->remote_size = remote_arg->area_size;
spin_lock_irqsave(&kcov_remote_lock, flags); for (i = 0; i < remote_arg->num_handles; i++) { if (!kcov_check_handle(remote_arg->handles[i], false, true, false)) {
spin_unlock_irqrestore(&kcov_remote_lock,
flags);
kcov_disable(t, kcov); return -EINVAL;
}
remote = kcov_remote_add(kcov, remote_arg->handles[i]); if (IS_ERR(remote)) {
spin_unlock_irqrestore(&kcov_remote_lock,
flags);
kcov_disable(t, kcov); return PTR_ERR(remote);
}
} if (remote_arg->common_handle) { if (!kcov_check_handle(remote_arg->common_handle, true, false, false)) {
spin_unlock_irqrestore(&kcov_remote_lock,
flags);
kcov_disable(t, kcov); return -EINVAL;
}
remote = kcov_remote_add(kcov,
remote_arg->common_handle); if (IS_ERR(remote)) {
spin_unlock_irqrestore(&kcov_remote_lock,
flags);
kcov_disable(t, kcov); return PTR_ERR(remote);
}
t->kcov_handle = remote_arg->common_handle;
}
spin_unlock_irqrestore(&kcov_remote_lock, flags); /* Put either in kcov_task_exit() or in KCOV_DISABLE. */
kcov_get(kcov); return 0; default: return -ENOTTY;
}
}
kcov = filep->private_data; switch (cmd) { case KCOV_INIT_TRACE: /* * Enable kcov in trace mode and setup buffer size. * Must happen before anything else. * * First check the size argument - it must be at least 2 * to hold the current position and one PC.
*/
size = arg; if (size < 2 || size > INT_MAX / sizeof(unsignedlong)) return -EINVAL;
area = vmalloc_user(size * sizeof(unsignedlong)); if (area == NULL) return -ENOMEM;
spin_lock_irqsave(&kcov->lock, flags); if (kcov->mode != KCOV_MODE_DISABLED) {
spin_unlock_irqrestore(&kcov->lock, flags);
vfree(area); return -EBUSY;
}
kcov->area = area;
kcov->size = size;
kcov->mode = KCOV_MODE_INIT;
spin_unlock_irqrestore(&kcov->lock, flags); return 0; case KCOV_REMOTE_ENABLE: if (get_user(remote_num_handles, (unsigned __user *)(arg +
offsetof(struct kcov_remote_arg, num_handles)))) return -EFAULT; if (remote_num_handles > KCOV_REMOTE_MAX_HANDLES) return -EINVAL;
remote_arg_size = struct_size(remote_arg, handles,
remote_num_handles);
remote_arg = memdup_user((void __user *)arg, remote_arg_size); if (IS_ERR(remote_arg)) return PTR_ERR(remote_arg); if (remote_arg->num_handles != remote_num_handles) {
kfree(remote_arg); return -EINVAL;
}
arg = (unsignedlong)remote_arg;
fallthrough; default: /* * All other commands can be normally executed under a spin lock, so we * obtain and release it here in order to simplify kcov_ioctl_locked().
*/
spin_lock_irqsave(&kcov->lock, flags);
res = kcov_ioctl_locked(kcov, cmd, arg);
spin_unlock_irqrestore(&kcov->lock, flags);
kfree(remote_arg); return res;
}
}
/* * kcov_remote_start() and kcov_remote_stop() can be used to annotate a section * of code in a kernel background thread or in a softirq to allow kcov to be * used to collect coverage from that part of code. * * The handle argument of kcov_remote_start() identifies a code section that is * used for coverage collection. A userspace process passes this handle to * KCOV_REMOTE_ENABLE ioctl to make the used kcov device start collecting * coverage for the code section identified by this handle. * * The usage of these annotations in the kernel code is different depending on * the type of the kernel thread whose code is being annotated. * * For global kernel threads that are spawned in a limited number of instances * (e.g. one USB hub_event() worker thread is spawned per USB HCD) and for * softirqs, each instance must be assigned a unique 4-byte instance id. The * instance id is then combined with a 1-byte subsystem id to get a handle via * kcov_remote_handle(subsystem_id, instance_id). * * For local kernel threads that are spawned from system calls handler when a * user interacts with some kernel interface (e.g. vhost workers), a handle is * passed from a userspace process as the common_handle field of the * kcov_remote_arg struct (note, that the user must generate a handle by using * kcov_remote_handle() with KCOV_SUBSYSTEM_COMMON as the subsystem id and an * arbitrary 4-byte non-zero number as the instance id). This common handle * then gets saved into the task_struct of the process that issued the * KCOV_REMOTE_ENABLE ioctl. When this process issues system calls that spawn * kernel threads, the common handle must be retrieved via kcov_common_handle() * and passed to the spawned threads via custom annotations. Those kernel * threads must in turn be annotated with kcov_remote_start(common_handle) and * kcov_remote_stop(). All of the threads that are spawned by the same process * obtain the same handle, hence the name "common". * * See Documentation/dev-tools/kcov.rst for more details. * * Internally, kcov_remote_start() looks up the kcov device associated with the * provided handle, allocates an area for coverage collection, and saves the * pointers to kcov and area into the current task_struct to allow coverage to * be collected via __sanitizer_cov_trace_pc(). * In turns kcov_remote_stop() clears those pointers from task_struct to stop * collecting coverage and copies all collected coverage into the kcov area.
*/
/* * Check that kcov_remote_start() is not called twice in background * threads nor called by user tasks (with enabled kcov).
*/
mode = READ_ONCE(t->kcov_mode); if (WARN_ON(in_task() && kcov_mode_enabled(mode))) {
local_unlock_irqrestore(&kcov_percpu_data.lock, flags); return;
} /* * Check that kcov_remote_start() is not called twice in softirqs. * Note, that kcov_remote_start() can be called from a softirq that * happened while collecting coverage from a background thread.
*/ if (WARN_ON(in_serving_softirq() && t->kcov_softirq)) {
local_unlock_irqrestore(&kcov_percpu_data.lock, flags); return;
}
spin_lock(&kcov_remote_lock);
remote = kcov_remote_find(handle); if (!remote) {
spin_unlock(&kcov_remote_lock);
local_unlock_irqrestore(&kcov_percpu_data.lock, flags); return;
}
kcov_debug("handle = %llx, context: %s\n", handle,
in_task() ? "task" : "softirq");
kcov = remote->kcov; /* Put in kcov_remote_stop(). */
kcov_get(kcov); /* * Read kcov fields before unlock to prevent races with * KCOV_DISABLE / kcov_remote_reset().
*/
mode = kcov->mode;
sequence = kcov->sequence; if (in_task()) {
size = kcov->remote_size;
area = kcov_remote_area_get(size);
} else {
size = CONFIG_KCOV_IRQ_AREA_SIZE;
area = this_cpu_ptr(&kcov_percpu_data)->irq_area;
}
spin_unlock(&kcov_remote_lock);
/* Can only happen when in_task(). */ if (!area) {
local_unlock_irqrestore(&kcov_percpu_data.lock, flags);
area = vmalloc(size * sizeof(unsignedlong)); if (!area) {
kcov_put(kcov); return;
}
local_lock_irqsave(&kcov_percpu_data.lock, flags);
}
mode = READ_ONCE(t->kcov_mode);
barrier(); if (!kcov_mode_enabled(mode)) {
local_unlock_irqrestore(&kcov_percpu_data.lock, flags); return;
} /* * When in softirq, check if the corresponding kcov_remote_start() * actually found the remote handle and started collecting coverage.
*/ if (in_serving_softirq() && !t->kcov_softirq) {
local_unlock_irqrestore(&kcov_percpu_data.lock, flags); return;
} /* Make sure that kcov_softirq is only set when in softirq. */ if (WARN_ON(!in_serving_softirq() && t->kcov_softirq)) {
local_unlock_irqrestore(&kcov_percpu_data.lock, flags); return;
}
kcov_stop(t); if (in_serving_softirq()) {
t->kcov_softirq = 0;
kcov_remote_softirq_stop(t);
}
spin_lock(&kcov->lock); /* * KCOV_DISABLE could have been called between kcov_remote_start() * and kcov_remote_stop(), hence the sequence check.
*/ if (sequence == kcov->sequence && kcov->remote)
kcov_move_area(kcov->mode, kcov->area, kcov->size, area);
spin_unlock(&kcov->lock);
if (in_task()) {
spin_lock(&kcov_remote_lock);
kcov_remote_area_put(area, size);
spin_unlock(&kcov_remote_lock);
}
/* Get in kcov_remote_start(). */
kcov_put(kcov);
}
EXPORT_SYMBOL(kcov_remote_stop);
/* See the comment before kcov_remote_start() for usage details. */
u64 kcov_common_handle(void)
{ if (!in_task()) return 0; return current->kcov_handle;
}
EXPORT_SYMBOL(kcov_common_handle);
pr_err("running self test\n"); /* * Test that interrupts don't produce spurious coverage. * The coverage callback filters out interrupt code, but only * after the handler updates preempt count. Some code periodically * leaks out of that section and leads to spurious coverage. * It's hard to call the actual interrupt handler directly, * so we just loop here for a bit waiting for a timer interrupt. * We set kcov_mode to enable tracing, but don't setup the area, * so any attempt to trace will crash. Note: we must not call any * potentially traced functions in this region.
*/
start = jiffies;
current->kcov_mode = KCOV_MODE_TRACE_PC; while ((jiffies - start) * MSEC_PER_SEC / HZ < 300)
;
current->kcov_mode = 0;
pr_err("done running self test\n");
} #endif
/* * The kcov debugfs file won't ever get removed and thus, * there is no need to protect it against removal races. The * use of debugfs_create_file_unsafe() is actually safe here.
*/
debugfs_create_file_unsafe("kcov", 0600, NULL, NULL, &kcov_fops);
#ifdef CONFIG_KCOV_SELFTEST
selftest(); #endif
return 0;
}
device_initcall(kcov_init);
¤ Dauer der Verarbeitung: 0.20 Sekunden
(vorverarbeitet)
¤
Die Informationen auf dieser Webseite wurden
nach bestem Wissen sorgfältig zusammengestellt. Es wird jedoch weder Vollständigkeit, noch Richtigkeit,
noch Qualität der bereit gestellten Informationen zugesichert.
Bemerkung:
Die farbliche Syntaxdarstellung ist noch experimentell.