/** * struct vmw_validation_bo_node - Buffer object validation metadata. * @base: Metadata used for TTM reservation- and validation. * @hash: A hash entry used for the duplicate detection hash table. * @coherent_count: If switching backup buffers, number of new coherent * resources that will have this buffer as a backup buffer. * * Bit fields are used since these structures are allocated and freed in * large numbers and space conservation is desired.
*/ struct vmw_validation_bo_node { struct ttm_validate_buffer base; struct vmwgfx_hash_item hash; unsignedint coherent_count;
}; /** * struct vmw_validation_res_node - Resource validation metadata. * @head: List head for the resource validation list. * @hash: A hash entry used for the duplicate detection hash table. * @res: Reference counted resource pointer. * @new_guest_memory_bo: Non ref-counted pointer to new guest memory buffer * to be assigned to a resource. * @new_guest_memory_offset: Offset into the new backup mob for resources * that can share MOBs. * @no_buffer_needed: Kernel does not need to allocate a MOB during validation, * the command stream provides a mob bind operation. * @switching_guest_memory_bo: The validation process is switching backup MOB. * @first_usage: True iff the resource has been seen only once in the current * validation batch. * @reserved: Whether the resource is currently reserved by this process. * @dirty_set: Change dirty status of the resource. * @dirty: Dirty information VMW_RES_DIRTY_XX. * @private: Optionally additional memory for caller-private data. * * Bit fields are used since these structures are allocated and freed in * large numbers and space conservation is desired.
*/ struct vmw_validation_res_node { struct list_head head; struct vmwgfx_hash_item hash; struct vmw_resource *res; struct vmw_bo *new_guest_memory_bo; unsignedlong new_guest_memory_offset;
u32 no_buffer_needed : 1;
u32 switching_guest_memory_bo : 1;
u32 first_usage : 1;
u32 reserved : 1;
u32 dirty : 1;
u32 dirty_set : 1; unsignedlongprivate[];
};
/** * vmw_validation_mem_alloc - Allocate kernel memory from the validation * context based allocator * @ctx: The validation context * @size: The number of bytes to allocated. * * The memory allocated may not exceed PAGE_SIZE, and the returned * address is aligned to sizeof(long). All memory allocated this way is * reclaimed after validation when calling any of the exported functions: * vmw_validation_unref_lists() * vmw_validation_revert() * vmw_validation_done() * * Return: Pointer to the allocated memory on success. NULL on failure.
*/ void *vmw_validation_mem_alloc(struct vmw_validation_context *ctx, unsignedint size)
{ void *addr;
size = vmw_validation_align(size); if (size > PAGE_SIZE) return NULL;
if (ctx->mem_size_left < size) { struct page *page = alloc_page(GFP_KERNEL | __GFP_ZERO); if (!page) return NULL;
/** * vmw_validation_mem_free - Free all memory allocated using * vmw_validation_mem_alloc() * @ctx: The validation context * * All memory previously allocated for this context using * vmw_validation_mem_alloc() is freed.
*/ staticvoid vmw_validation_mem_free(struct vmw_validation_context *ctx)
{ struct page *entry, *next;
/** * vmw_validation_add_resource - Add a resource to the validation context. * @ctx: The validation context. * @res: The resource. * @priv_size: Size of private, additional metadata. * @dirty: Whether to change dirty status. * @p_node: Output pointer of additional metadata address. * @first_usage: Whether this was the first time this resource was seen. * * Return: Zero on success, negative error code otherwise.
*/ int vmw_validation_add_resource(struct vmw_validation_context *ctx, struct vmw_resource *res,
size_t priv_size,
u32 dirty, void **p_node, bool *first_usage)
{ struct vmw_validation_res_node *node;
node->first_usage = 1; if (!res->dev_priv->has_mob) {
list_add_tail(&node->head, &ctx->resource_list);
} else { switch (vmw_res_type(res)) { case vmw_res_context: case vmw_res_dx_context:
list_add(&node->head, &ctx->resource_ctx_list); break; case vmw_res_cotable:
list_add_tail(&node->head, &ctx->resource_ctx_list); break; default:
list_add_tail(&node->head, &ctx->resource_list); break;
}
}
out_fill: if (dirty) {
node->dirty_set = 1; /* Overwriting previous information here is intentional! */
node->dirty = (dirty & VMW_RES_DIRTY_SET) ? 1 : 0;
} if (first_usage)
*first_usage = node->first_usage; if (p_node)
*p_node = &node->private;
return 0;
}
/** * vmw_validation_res_set_dirty - Register a resource dirty set or clear during * validation. * @ctx: The validation context. * @val_private: The additional meta-data pointer returned when the * resource was registered with the validation context. Used to identify * the resource. * @dirty: Dirty information VMW_RES_DIRTY_XX
*/ void vmw_validation_res_set_dirty(struct vmw_validation_context *ctx, void *val_private, u32 dirty)
{ struct vmw_validation_res_node *val;
if (!dirty) return;
val = container_of(val_private, typeof(*val), private);
val->dirty_set = 1; /* Overwriting previous information here is intentional! */
val->dirty = (dirty & VMW_RES_DIRTY_SET) ? 1 : 0;
}
/** * vmw_validation_res_switch_backup - Register a backup MOB switch during * validation. * @ctx: The validation context. * @val_private: The additional meta-data pointer returned when the * resource was registered with the validation context. Used to identify * the resource. * @vbo: The new backup buffer object MOB. This buffer object needs to have * already been registered with the validation context. * @guest_memory_offset: Offset into the new backup MOB.
*/ void vmw_validation_res_switch_backup(struct vmw_validation_context *ctx, void *val_private, struct vmw_bo *vbo, unsignedlong guest_memory_offset)
{ struct vmw_validation_res_node *val;
val = container_of(val_private, typeof(*val), private);
val->switching_guest_memory_bo = 1; if (val->first_usage)
val->no_buffer_needed = 1;
/** * vmw_validation_res_reserve - Reserve all resources registered with this * validation context. * @ctx: The validation context. * @intr: Use interruptible waits when possible. * * Return: Zero on success, -ERESTARTSYS if interrupted. Negative error * code on failure.
*/ int vmw_validation_res_reserve(struct vmw_validation_context *ctx, bool intr)
{ struct vmw_validation_res_node *val; int ret = 0;
/** * vmw_validation_res_unreserve - Unreserve all reserved resources * registered with this validation context. * @ctx: The validation context. * @backoff: Whether this is a backoff- of a commit-type operation. This * is used to determine whether to switch backup MOBs or not.
*/ void vmw_validation_res_unreserve(struct vmw_validation_context *ctx, bool backoff)
{ struct vmw_validation_res_node *val;
/** * vmw_validation_bo_validate - Validate all buffer objects registered with * the validation context. * @ctx: The validation context. * @intr: Whether to perform waits interruptible if possible. * * Return: Zero on success, -ERESTARTSYS if interrupted, * negative error code on failure.
*/ int vmw_validation_bo_validate(struct vmw_validation_context *ctx, bool intr)
{ struct vmw_validation_bo_node *entry; int ret;
ret = vmw_validation_bo_validate_single(entry->base.bo, intr);
if (ret) return ret;
/* * Rather than having the resource code allocating the bo * dirty tracker in resource_unreserve() where we can't fail, * Do it here when validating the buffer object.
*/ if (entry->coherent_count) { unsignedint coherent_count = entry->coherent_count;
while (coherent_count) {
ret = vmw_bo_dirty_add(vbo); if (ret) return ret;
if (vbo->dirty)
vmw_bo_dirty_scan(vbo);
} return 0;
}
/** * vmw_validation_res_validate - Validate all resources registered with the * validation context. * @ctx: The validation context. * @intr: Whether to perform waits interruptible if possible. * * Before this function is called, all resource backup buffers must have * been validated. * * Return: Zero on success, -ERESTARTSYS if interrupted, * negative error code on failure.
*/ int vmw_validation_res_validate(struct vmw_validation_context *ctx, bool intr)
{ struct vmw_validation_res_node *val; int ret;
ret = vmw_resource_validate(res, intr, val->dirty_set &&
val->dirty); if (ret) { if (ret != -ERESTARTSYS)
DRM_ERROR("Failed to validate resource.\n"); return ret;
}
/* Check if the resource switched backup buffer */ if (backup && res->guest_memory_bo && backup != res->guest_memory_bo) { struct vmw_bo *vbo = res->guest_memory_bo;
vmw_bo_placement_set(vbo, res->func->domain,
res->func->busy_domain);
ret = vmw_validation_add_bo(ctx, vbo); if (ret) return ret;
}
} return 0;
}
/** * vmw_validation_drop_ht - Reset the hash table used for duplicate finding * and unregister it from this validation context. * @ctx: The validation context. * * The hash table used for duplicate finding is an expensive resource and * may be protected by mutexes that may cause deadlocks during resource * unreferencing if held. After resource- and buffer object registering, * there is no longer any use for this hash table, so allow freeing it * either to shorten any mutex locking time, or before resources- and * buffer objects are freed during validation context cleanup.
*/ void vmw_validation_drop_ht(struct vmw_validation_context *ctx)
{ struct vmw_validation_bo_node *entry; struct vmw_validation_res_node *val;
/** * vmw_validation_unref_lists - Unregister previously registered buffer * object and resources. * @ctx: The validation context. * * Note that this function may cause buffer object- and resource destructors * to be invoked.
*/ void vmw_validation_unref_lists(struct vmw_validation_context *ctx)
{ struct vmw_validation_bo_node *entry; struct vmw_validation_res_node *val;
/* * No need to detach each list entry since they are all freed with * vmw_validation_free_mem. Just make the inaccessible.
*/
INIT_LIST_HEAD(&ctx->bo_list);
INIT_LIST_HEAD(&ctx->resource_list);
vmw_validation_mem_free(ctx);
}
/** * vmw_validation_prepare - Prepare a validation context for command * submission. * @ctx: The validation context. * @mutex: The mutex used to protect resource reservation. * @intr: Whether to perform waits interruptible if possible. * * Note that the single reservation mutex @mutex is an unfortunate * construct. Ideally resource reservation should be moved to per-resource * ww_mutexes. * If this functions doesn't return Zero to indicate success, all resources * are left unreserved but still referenced. * Return: Zero on success, -ERESTARTSYS if interrupted, negative error code * on error.
*/ int vmw_validation_prepare(struct vmw_validation_context *ctx, struct mutex *mutex, bool intr)
{ int ret = 0;
if (mutex) { if (intr)
ret = mutex_lock_interruptible(mutex); else
mutex_lock(mutex); if (ret) return -ERESTARTSYS;
}
ctx->res_mutex = mutex;
ret = vmw_validation_res_reserve(ctx, intr); if (ret) goto out_no_res_reserve;
ret = vmw_validation_bo_reserve(ctx, intr); if (ret) goto out_no_bo_reserve;
ret = vmw_validation_bo_validate(ctx, intr); if (ret) goto out_no_validate;
ret = vmw_validation_res_validate(ctx, intr); if (ret) goto out_no_validate;
return 0;
out_no_validate:
vmw_validation_bo_backoff(ctx);
out_no_bo_reserve:
vmw_validation_res_unreserve(ctx, true);
out_no_res_reserve: if (mutex)
mutex_unlock(mutex);
return ret;
}
/** * vmw_validation_revert - Revert validation actions if command submission * failed. * * @ctx: The validation context. * * The caller still needs to unref resources after a call to this function.
*/ void vmw_validation_revert(struct vmw_validation_context *ctx)
{
vmw_validation_bo_backoff(ctx);
vmw_validation_res_unreserve(ctx, true); if (ctx->res_mutex)
mutex_unlock(ctx->res_mutex);
vmw_validation_unref_lists(ctx);
}
/** * vmw_validation_done - Commit validation actions after command submission * success. * @ctx: The validation context. * @fence: Fence with which to fence all buffer objects taking part in the * command submission. * * The caller does NOT need to unref resources after a call to this function.
*/ void vmw_validation_done(struct vmw_validation_context *ctx, struct vmw_fence_obj *fence)
{
vmw_validation_bo_fence(ctx, fence);
vmw_validation_res_unreserve(ctx, false); if (ctx->res_mutex)
mutex_unlock(ctx->res_mutex);
vmw_validation_unref_lists(ctx);
}
/** * vmw_validation_preload_bo - Preload the validation memory allocator for a * call to vmw_validation_add_bo(). * @ctx: Pointer to the validation context. * * Iff this function returns successfully, the next call to * vmw_validation_add_bo() is guaranteed not to sleep. An error is not fatal * but voids the guarantee. * * Returns: Zero if successful, %-EINVAL otherwise.
*/ int vmw_validation_preload_bo(struct vmw_validation_context *ctx)
{ unsignedint size = sizeof(struct vmw_validation_bo_node);
if (!vmw_validation_mem_alloc(ctx, size)) return -ENOMEM;
ctx->mem_size_left += size; return 0;
}
/** * vmw_validation_preload_res - Preload the validation memory allocator for a * call to vmw_validation_add_res(). * @ctx: Pointer to the validation context. * @size: Size of the validation node extra data. See below. * * Iff this function returns successfully, the next call to * vmw_validation_add_res() with the same or smaller @size is guaranteed not to * sleep. An error is not fatal but voids the guarantee. * * Returns: Zero if successful, %-EINVAL otherwise.
*/ int vmw_validation_preload_res(struct vmw_validation_context *ctx, unsignedint size)
{
size = vmw_validation_align(sizeof(struct vmw_validation_res_node) +
size) +
vmw_validation_align(sizeof(struct vmw_validation_bo_node)); if (!vmw_validation_mem_alloc(ctx, size)) return -ENOMEM;
ctx->mem_size_left += size; return 0;
}
/** * vmw_validation_bo_backoff - Unreserve buffer objects registered with a * validation context * @ctx: The validation context * * This function unreserves the buffer objects previously reserved using * vmw_validation_bo_reserve. It's typically used as part of an error path
*/ void vmw_validation_bo_backoff(struct vmw_validation_context *ctx)
{ struct vmw_validation_bo_node *entry;
Die Informationen auf dieser Webseite wurden
nach bestem Wissen sorgfältig zusammengestellt. Es wird jedoch weder Vollständigkeit, noch Richtigkeit,
noch Qualität der bereit gestellten Informationen zugesichert.
Bemerkung:
Die farbliche Syntaxdarstellung und die Messung sind noch experimentell.