/* * We defer making "oops" entries appear in pstore - see * whether the system is actually still running well enough * to let someone see the entry
*/ staticint pstore_update_ms = -1;
module_param_named(update_ms, pstore_update_ms, int, 0600);
MODULE_PARM_DESC(update_ms, "milliseconds before pstore updates its content " "(default is -1, which means runtime updates are disabled; " "enabling this option may not be safe; it may lead to further " "corruption on Oopses)");
/* Names should be in the same order as the enum pstore_type_id */ staticconstchar * const pstore_type_names[] = { "dmesg", "mce", "console", "ftrace", "rtas", "powerpc-ofw", "powerpc-common", "pmsg", "powerpc-opal",
};
/* * psinfo_lock protects "psinfo" during calls to * pstore_register(), pstore_unregister(), and * the filesystem mount/unmount routines.
*/ static DEFINE_MUTEX(psinfo_lock); struct pstore_info *psinfo;
staticchar *backend;
module_param(backend, charp, 0444);
MODULE_PARM_DESC(backend, "specific backend to use");
/* * pstore no longer implements compression via the crypto API, and only * supports zlib deflate compression implemented using the zlib library * interface. This removes additional complexity which is hard to justify for a * diagnostic facility that has to operate in conditions where the system may * have become unstable. Zlib deflate is comparatively small in terms of code * size, and compresses ASCII text comparatively well. In terms of compression * speed, deflate is not the best performer but for recording the log output on * a kernel panic, this is not considered critical. * * The only remaining arguments supported by the compress= module parameter are * 'deflate' and 'none'. To retain compatibility with existing installations, * all other values are logged and replaced with 'deflate'.
*/ staticchar *compress = "deflate";
module_param(compress, charp, 0444);
MODULE_PARM_DESC(compress, "compression to use");
/* How much of the kernel log to snapshot */ unsignedint kmsg_bytes = CONFIG_PSTORE_DEFAULT_KMSG_BYTES;
module_param(kmsg_bytes, uint, 0444);
MODULE_PARM_DESC(kmsg_bytes, "amount of kernel log to snapshot (in bytes)");
staticvoid *compress_workspace;
/* * Compression is only used for dmesg output, which consists of low-entropy * ASCII text, and so we can assume worst-case 60%.
*/ #define DMESG_COMP_PERCENT 60
staticbool pstore_cannot_block_path(enum kmsg_dump_reason reason)
{ /* * In case of NMI path, pstore shouldn't be blocked * regardless of reason.
*/ if (in_nmi()) returntrue;
switch (reason) { /* In panic case, other cpus are stopped by smp_send_stop(). */ case KMSG_DUMP_PANIC: /* * Emergency restart shouldn't be blocked by spinning on * pstore_info::buf_lock.
*/ case KMSG_DUMP_EMERG: returntrue; default: returnfalse;
}
}
/* Skip if not built-in or compression disabled. */ if (!IS_ENABLED(CONFIG_PSTORE_COMPRESS) || !compress ||
!strcmp(compress, "none")) {
compress = NULL; return;
}
if (strcmp(compress, "deflate")) {
pr_err("Unsupported compression '%s', falling back to deflate\n",
compress);
compress = "deflate";
}
/* * The compression buffer only needs to be as large as the maximum * uncompressed record size, since any record that would be expanded by * compression is just stored uncompressed.
*/
compressed_size = (psinfo->bufsize * 100) / DMESG_COMP_PERCENT;
buf = kvzalloc(compressed_size, GFP_KERNEL); if (!buf) {
pr_err("Failed %zu byte compression buffer allocation for: %s\n",
psinfo->bufsize, compress); return;
}
compress_workspace =
vmalloc(zlib_deflate_workspacesize(MAX_WBITS, DEF_MEM_LEVEL)); if (!compress_workspace) {
pr_err("Failed to allocate zlib deflate workspace\n");
kvfree(buf); return;
}
/* A non-NULL big_oops_buf indicates compression is available. */
big_oops_buf = buf;
max_compressed_size = compressed_size;
/* Report zeroed timestamp if called before timekeeping has resumed. */
record->time = ns_to_timespec64(ktime_get_real_fast_ns());
}
/* * callback from kmsg_dump. Save as much as we can (up to kmsg_bytes) from the * end of the buffer.
*/ staticvoid pstore_dump(struct kmsg_dumper *dumper, struct kmsg_dump_detail *detail)
{ struct kmsg_dump_iter iter; unsignedint remaining = READ_ONCE(kmsg_bytes); unsignedlong total = 0; constchar *why; unsignedint part = 1; unsignedlong flags = 0; int saved_ret = 0; int ret;
why = kmsg_dump_reason_str(detail->reason);
if (pstore_cannot_block_path(detail->reason)) { if (!raw_spin_trylock_irqsave(&psinfo->buf_lock, flags)) {
pr_err("dump skipped in %s path because of concurrent dump\n",
in_nmi() ? "NMI" : why); return;
}
} else {
raw_spin_lock_irqsave(&psinfo->buf_lock, flags);
}
kmsg_dump_rewind(&iter);
oopscount++; while (total < remaining) { char *dst;
size_t dst_size; int header_size; int zipped_len = -1;
size_t dump_size; struct pstore_record record;
if (zipped_len > 0) {
record.compressed = true;
record.size = zipped_len;
} else { /* * Compression failed, so the buffer is most * likely filled with binary data that does not * compress as well as ASCII text. Copy as much * of the uncompressed data as possible into * the pstore record, and discard the rest.
*/
record.size = psinfo->bufsize;
memcpy(psinfo->buf, dst, psinfo->bufsize);
}
} else {
record.size = header_size + dump_size;
}
ret = psinfo->write(&record); if (ret == 0 && detail->reason == KMSG_DUMP_OOPS) {
pstore_new_entry = 1;
pstore_timer_kick();
} else { /* Preserve only the first non-zero returned value. */ if (!saved_ret)
saved_ret = ret;
}
total += record.size;
part++;
}
raw_spin_unlock_irqrestore(&psinfo->buf_lock, flags);
/* * Register with kmsg_dump to save last part of console log on panic.
*/ staticvoid pstore_register_kmsg(void)
{
kmsg_dump_register(&pstore_dumper);
}
staticvoid pstore_register_console(void)
{ /* Show which backend is going to get console writes. */
strscpy(pstore_console.name, psinfo->name, sizeof(pstore_console.name)); /* * Always initialize flags here since prior unregister_console() * calls may have changed settings (specifically CON_ENABLED).
*/
pstore_console.flags = CON_PRINTBUFFER | CON_ENABLED | CON_ANYTIME;
register_console(&pstore_console);
}
staticint pstore_write_user_compat(struct pstore_record *record, constchar __user *buf)
{ int ret = 0;
if (record->buf) return -EINVAL;
record->buf = vmemdup_user(buf, record->size); if (IS_ERR(record->buf)) {
ret = PTR_ERR(record->buf); goto out;
}
ret = record->psi->write(record);
kvfree(record->buf);
out:
record->buf = NULL;
return unlikely(ret < 0) ? ret : record->size;
}
/* * platform specific persistent storage driver registers with * us here. If pstore is already mounted, call the platform * read function right away to populate the file system. If not * then the pstore mount code will call us later to fill out * the file system.
*/ int pstore_register(struct pstore_info *psi)
{ char *new_backend;
if (backend && strcmp(backend, psi->name)) {
pr_warn("backend '%s' already in use: ignoring '%s'\n",
backend, psi->name); return -EBUSY;
}
/* Sanity check flags. */ if (!psi->flags) {
pr_warn("backend '%s' must support at least one frontend\n",
psi->name); return -EINVAL;
}
/* Check for required functions. */ if (!psi->read || !psi->write) {
pr_warn("backend '%s' must implement read() and write()\n",
psi->name); return -EINVAL;
}
new_backend = kstrdup(psi->name, GFP_KERNEL); if (!new_backend) return -ENOMEM;
void pstore_unregister(struct pstore_info *psi)
{ /* It's okay to unregister nothing. */ if (!psi) return;
mutex_lock(&psinfo_lock);
/* Only one backend can be registered at a time. */ if (WARN_ON(psi != psinfo)) {
mutex_unlock(&psinfo_lock); return;
}
/* Unregister all callbacks. */ if (psi->flags & PSTORE_FLAGS_PMSG)
pstore_unregister_pmsg(); if (psi->flags & PSTORE_FLAGS_FTRACE)
pstore_unregister_ftrace(); if (psi->flags & PSTORE_FLAGS_CONSOLE)
pstore_unregister_console(); if (psi->flags & PSTORE_FLAGS_DMESG)
pstore_unregister_kmsg();
/* Stop timer and make sure all work has finished. */
timer_delete_sync(&pstore_timer);
flush_work(&pstore_work);
/* Remove all backend records from filesystem tree. */
pstore_put_backend_records(psi);
free_buf_for_compression();
psinfo = NULL;
kfree(backend);
backend = NULL;
pr_info("Unregistered %s as persistent store backend\n", psi->name);
mutex_unlock(&psinfo_lock);
}
EXPORT_SYMBOL_GPL(pstore_unregister);
staticvoid decompress_record(struct pstore_record *record, struct z_stream_s *zstream)
{ int ret; int unzipped_len; char *unzipped, *workspace;
size_t max_uncompressed_size;
if (!IS_ENABLED(CONFIG_PSTORE_COMPRESS) || !record->compressed) return;
/* Only PSTORE_TYPE_DMESG support compression. */ if (record->type != PSTORE_TYPE_DMESG) {
pr_warn("ignored compressed record type %d\n", record->type); return;
}
/* Missing compression buffer means compression was not initialized. */ if (!zstream->workspace) {
pr_warn("no decompression method initialized!\n"); return;
}
ret = zlib_inflateReset(zstream); if (ret != Z_OK) {
pr_err("zlib_inflateReset() failed, ret = %d!\n", ret); return;
}
/* Allocate enough space to hold max decompression and ECC. */
max_uncompressed_size = 3 * psinfo->bufsize;
workspace = kvzalloc(max_uncompressed_size + record->ecc_notice_size,
GFP_KERNEL); if (!workspace) return;
/* Copy decompressed contents into an minimum-sized allocation. */
unzipped = kvmemdup(workspace, unzipped_len + record->ecc_notice_size,
GFP_KERNEL);
kvfree(workspace); if (!unzipped) return;
/* Swap out compressed contents with decompressed contents. */
kvfree(record->buf);
record->buf = unzipped;
record->size = unzipped_len;
record->compressed = false;
}
/* * Read all the records from one persistent store backend. Create * files in our filesystem. Don't warn about -EEXIST errors * when we are re-scanning the backing store looking to add new * error records.
*/ void pstore_get_backend_records(struct pstore_info *psi, struct dentry *root, int quiet)
{ int failed = 0; unsignedint stop_loop = 65536; struct z_stream_s zstream = {};
mutex_lock(&psi->read_mutex); if (psi->open && psi->open(psi)) goto out;
/* * Backend callback read() allocates record.buf. decompress_record() * may reallocate record.buf. On success, pstore_mkfile() will keep * the record.buf, so free it only on failure.
*/ for (; stop_loop; stop_loop--) { struct pstore_record *record; int rc;
record = kzalloc(sizeof(*record), GFP_KERNEL); if (!record) {
pr_err("out of memory creating record\n"); break;
}
pstore_record_init(record, psi);
record->size = psi->read(record);
/* No more records left in backend? */ if (record->size <= 0) {
kfree(record); break;
}
decompress_record(record, &zstream);
rc = pstore_mkfile(root, record); if (rc) { /* pstore_mkfile() did not take record, so free it. */
kvfree(record->buf);
kfree(record->priv);
kfree(record); if (rc != -EEXIST || !quiet)
failed++;
}
} if (psi->close)
psi->close(psi);
out:
mutex_unlock(&psi->read_mutex);
if (IS_ENABLED(CONFIG_PSTORE_COMPRESS) && compress) { if (zlib_inflateEnd(&zstream) != Z_OK)
pr_warn("zlib_inflateEnd() failed\n");
kvfree(zstream.workspace);
}
if (failed)
pr_warn("failed to create %d record(s) from '%s'\n",
failed, psi->name); if (!stop_loop)
pr_err("looping? Too many records seen from '%s'\n",
psi->name);
}
Die Informationen auf dieser Webseite wurden
nach bestem Wissen sorgfältig zusammengestellt. Es wird jedoch weder Vollständigkeit, noch Richtigkeit,
noch Qualität der bereit gestellten Informationen zugesichert.
Bemerkung:
Die farbliche Syntaxdarstellung und die Messung sind noch experimentell.