// SPDX-License-Identifier: GPL-2.0-or-later /* * Squashfs - a compressed read only filesystem for Linux * * Copyright (c) 2002, 2003, 2004, 2005, 2006, 2007, 2008 * Phillip Lougher <phillip@squashfs.org.uk> * * cache.c
*/
/* * Blocks in Squashfs are compressed. To avoid repeatedly decompressing * recently accessed data Squashfs uses two small metadata and fragment caches. * * This file implements a generic cache implementation used for both caches, * plus functions layered ontop of the generic cache implementation to * access the metadata and fragment caches. * * To avoid out of memory and fragmentation issues with vmalloc the cache * uses sequences of kmalloced PAGE_SIZE buffers. * * It should be noted that the cache is not used for file datablocks, these * are decompressed and cached in the page-cache in the normal way. The * cache is only used to temporarily cache fragment and metadata blocks * which have been read as as a result of a metadata (i.e. inode or * directory) or fragment access. Because metadata and fragments are packed * together into blocks (to gain greater compression) the read of a particular * piece of metadata or fragment will retrieve other metadata/fragments which * have been packed with it, these because of locality-of-reference may be read * in the near future. Temporarily caching them ensures they are available for * near future access without requiring an additional read and decompress.
*/
/* * Look-up block in cache, and increment usage count. If not in cache, read * and decompress it from disk.
*/ struct squashfs_cache_entry *squashfs_cache_get(struct super_block *sb, struct squashfs_cache *cache, u64 block, int length)
{ int i, n; struct squashfs_cache_entry *entry;
spin_lock(&cache->lock);
while (1) { for (i = cache->curr_blk, n = 0; n < cache->entries; n++) { if (cache->entry[i].block == block) {
cache->curr_blk = i; break;
}
i = (i + 1) % cache->entries;
}
if (n == cache->entries) { /* * Block not in cache, if all cache entries are used * go to sleep waiting for one to become available.
*/ if (cache->unused == 0) {
cache->num_waiters++;
spin_unlock(&cache->lock);
wait_event(cache->wait_queue, cache->unused);
spin_lock(&cache->lock);
cache->num_waiters--; continue;
}
/* * At least one unused cache entry. A simple * round-robin strategy is used to choose the entry to * be evicted from the cache.
*/
i = cache->next_blk; for (n = 0; n < cache->entries; n++) { if (cache->entry[i].refcount == 0) break;
i = (i + 1) % cache->entries;
}
cache->next_blk = (i + 1) % cache->entries;
entry = &cache->entry[i];
/* * Initialise chosen cache entry, and fill it in from * disk.
*/
cache->unused--;
entry->block = block;
entry->refcount = 1;
entry->pending = 1;
entry->num_waiters = 0;
entry->error = 0;
spin_unlock(&cache->lock);
if (entry->length < 0)
entry->error = entry->length;
entry->pending = 0;
/* * While filling this entry one or more other processes * have looked it up in the cache, and have slept * waiting for it to become available.
*/ if (entry->num_waiters) {
spin_unlock(&cache->lock);
wake_up_all(&entry->wait_queue);
} else
spin_unlock(&cache->lock);
goto out;
}
/* * Block already in cache. Increment refcount so it doesn't * get reused until we're finished with it, if it was * previously unused there's one less cache entry available * for reuse.
*/
entry = &cache->entry[i]; if (entry->refcount == 0)
cache->unused--;
entry->refcount++;
/* * If the entry is currently being filled in by another process * go to sleep waiting for it to become available.
*/ if (entry->pending) {
entry->num_waiters++;
spin_unlock(&cache->lock);
wait_event(entry->wait_queue, !entry->pending);
} else
spin_unlock(&cache->lock);
if (entry->error)
ERROR("Unable to read %s cache entry [%llx]\n", cache->name,
block); return entry;
}
/* * Release cache entry, once usage count is zero it can be reused.
*/ void squashfs_cache_put(struct squashfs_cache_entry *entry)
{ struct squashfs_cache *cache = entry->cache;
spin_lock(&cache->lock);
entry->refcount--; if (entry->refcount == 0) {
cache->unused++; /* * If there's any processes waiting for a block to become * available, wake one up.
*/ if (cache->num_waiters) {
spin_unlock(&cache->lock);
wake_up(&cache->wait_queue); return;
}
}
spin_unlock(&cache->lock);
}
/* * Delete cache reclaiming all kmalloced buffers.
*/ void squashfs_cache_delete(struct squashfs_cache *cache)
{ int i, j;
if (IS_ERR(cache) || cache == NULL) return;
for (i = 0; i < cache->entries; i++) { if (cache->entry[i].data) { for (j = 0; j < cache->pages; j++)
kfree(cache->entry[i].data[j]);
kfree(cache->entry[i].data);
}
kfree(cache->entry[i].actor);
}
kfree(cache->entry);
kfree(cache);
}
/* * Initialise cache allocating the specified number of entries, each of * size block_size. To avoid vmalloc fragmentation issues each entry * is allocated as a sequence of kmalloced PAGE_SIZE buffers.
*/ struct squashfs_cache *squashfs_cache_init(char *name, int entries, int block_size)
{ int i, j; struct squashfs_cache *cache;
if (entries == 0) return NULL;
cache = kzalloc(sizeof(*cache), GFP_KERNEL); if (cache == NULL) {
ERROR("Failed to allocate %s cache\n", name); return ERR_PTR(-ENOMEM);
}
cache->entry = kcalloc(entries, sizeof(*(cache->entry)), GFP_KERNEL); if (cache->entry == NULL) {
ERROR("Failed to allocate %s cache\n", name); goto cleanup;
}
/* * Copy up to length bytes from cache entry to buffer starting at offset bytes * into the cache entry. If there's not length bytes then copy the number of * bytes available. In all cases return the number of bytes copied.
*/ int squashfs_copy_data(void *buffer, struct squashfs_cache_entry *entry, int offset, int length)
{ int remaining = length;
/* * Read length bytes from metadata position <block, offset> (block is the * start of the compressed block on disk, and offset is the offset into * the block once decompressed). Data is packed into consecutive blocks, * and length bytes may require reading more than one block.
*/ int squashfs_read_metadata(struct super_block *sb, void *buffer,
u64 *block, int *offset, int length)
{ struct squashfs_sb_info *msblk = sb->s_fs_info; int bytes, res = length; struct squashfs_cache_entry *entry;
/* * Look-up in the fragmment cache the fragment located at <start_block> in the * filesystem. If necessary read and decompress it from disk.
*/ struct squashfs_cache_entry *squashfs_get_fragment(struct super_block *sb,
u64 start_block, int length)
{ struct squashfs_sb_info *msblk = sb->s_fs_info;
/* * Read and decompress the datablock located at <start_block> in the * filesystem. The cache is used here to avoid duplicating locking and * read/decompress code.
*/ struct squashfs_cache_entry *squashfs_get_datablock(struct super_block *sb,
u64 start_block, int length)
{ struct squashfs_sb_info *msblk = sb->s_fs_info;
Die Informationen auf dieser Webseite wurden
nach bestem Wissen sorgfältig zusammengestellt. Es wird jedoch weder Vollständigkeit, noch Richtigkeit,
noch Qualität der bereit gestellten Informationen zugesichert.
Bemerkung:
Die farbliche Syntaxdarstellung und die Messung sind noch experimentell.