/* * Sometimes we can't deal with a bio straight away. We put them in prison * where they can't cause any mischief. Bios are put in a cell identified * by a key, multiple bios can be in the same cell. When the cell is * subsequently unlocked the bios become available.
*/ struct dm_bio_prison;
/* * Keys define a range of blocks within either a virtual or physical * device.
*/ struct dm_cell_key { intvirtual;
dm_thin_id dev;
dm_block_t block_begin, block_end;
};
/* * The range of a key (block_end - block_begin) must not * exceed BIO_PRISON_MAX_RANGE. Also the range must not * cross a similarly sized boundary. * * Must be a power of 2.
*/ #define BIO_PRISON_MAX_RANGE 1024 #define BIO_PRISON_MAX_RANGE_SHIFT 10
/* * Treat this as opaque, only in header so callers can manage allocation * themselves.
*/ struct dm_bio_prison_cell { struct list_head user_list; /* for client use */ struct rb_node node;
struct dm_cell_key key; struct bio *holder; struct bio_list bios;
};
/* * These two functions just wrap a mempool. This is a transitory step: * Eventually all bio prison clients should manage their own cell memory. * * Like mempool_alloc(), dm_bio_prison_alloc_cell() can only fail if called * in interrupt context or passed GFP_NOWAIT.
*/ struct dm_bio_prison_cell *dm_bio_prison_alloc_cell(struct dm_bio_prison *prison,
gfp_t gfp); void dm_bio_prison_free_cell(struct dm_bio_prison *prison, struct dm_bio_prison_cell *cell);
/* * Returns false if key is beyond BIO_PRISON_MAX_RANGE or spans a boundary.
*/ bool dm_cell_key_has_valid_range(struct dm_cell_key *key);
/* * An atomic op that combines retrieving or creating a cell, and adding a * bio to it. * * Returns 1 if the cell was already held, 0 if @inmate is the new holder.
*/ int dm_bio_detain(struct dm_bio_prison *prison, struct dm_cell_key *key, struct bio *inmate, struct dm_bio_prison_cell *cell_prealloc, struct dm_bio_prison_cell **cell_result);
/* * Visits the cell and then releases. Guarantees no new inmates are * inserted between the visit and release.
*/ void dm_cell_visit_release(struct dm_bio_prison *prison, void (*visit_fn)(void *, struct dm_bio_prison_cell *), void *context, struct dm_bio_prison_cell *cell);
/* * We use the deferred set to keep track of pending reads to shared blocks. * We do this to ensure the new mapping caused by a write isn't performed * until these prior reads have completed. Otherwise the insertion of the * new mapping could free the old block that the read bios are mapped to.
*/
Die Informationen auf dieser Webseite wurden
nach bestem Wissen sorgfältig zusammengestellt. Es wird jedoch weder Vollständigkeit, noch Richtigkeit,
noch Qualität der bereit gestellten Informationen zugesichert.
Bemerkung:
Die farbliche Syntaxdarstellung und die Messung sind noch experimentell.