/** * struct vsp1_dl_ext_header - Extended display list header * @padding: padding zero bytes for alignment * @pre_ext_dl_num_cmd: number of pre-extended command bodies to parse * @flags: enables or disables execution of the pre and post command * @pre_ext_dl_plist: start address of pre-extended display list bodies * @post_ext_dl_num_cmd: number of post-extended command bodies to parse * @post_ext_dl_plist: start address of post-extended display list bodies
*/ struct vsp1_dl_ext_header {
u32 padding;
/* * The datasheet represents flags as stored before pre_ext_dl_num_cmd, * expecting 32-bit accesses. The flags are appropriate to the whole * header, not just the pre_ext command, and thus warrant being * separated out. Due to byte ordering, and representing as 16 bit * values here, the flags must be positioned after the * pre_ext_dl_num_cmd.
*/
u16 pre_ext_dl_num_cmd;
u16 flags;
u32 pre_ext_dl_plist;
/** * struct vsp1_pre_ext_dl_body - Pre Extended Display List Body * @opcode: Extended display list command operation code * @flags: Pre-extended command flags. These are specific to each command * @address_set: Source address set pointer. Must have 16-byte alignment * @reserved: Zero bits for alignment.
*/ struct vsp1_pre_ext_dl_body {
u32 opcode;
u32 flags;
u32 address_set;
u32 reserved;
} __packed;
/** * struct vsp1_dl_body - Display list body * @list: entry in the display list list of bodies * @free: entry in the pool free body list * @refcnt: reference tracking for the body * @pool: pool to which this body belongs * @entries: array of entries * @dma: DMA address of the entries * @size: size of the DMA memory in bytes * @num_entries: number of stored entries * @max_entries: number of entries available
*/ struct vsp1_dl_body { struct list_head list; struct list_head free;
/** * struct vsp1_dl_body_pool - display list body pool * @dma: DMA address of the entries * @size: size of the full DMA memory pool in bytes * @mem: CPU memory pointer for the pool * @bodies: Array of DLB structures for the pool * @free: List of free DLB entries * @lock: Protects the free list * @vsp1: the VSP1 device
*/ struct vsp1_dl_body_pool { /* DMA allocation */
dma_addr_t dma;
size_t size; void *mem;
/** * struct vsp1_dl_cmd_pool - Display List commands pool * @dma: DMA address of the entries * @size: size of the full DMA memory pool in bytes * @mem: CPU memory pointer for the pool * @cmds: Array of command structures for the pool * @free: Free pool entries * @lock: Protects the free list * @vsp1: the VSP1 device
*/ struct vsp1_dl_cmd_pool { /* DMA allocation */
dma_addr_t dma;
size_t size; void *mem;
/** * struct vsp1_dl_list - Display list * @list: entry in the display list manager lists * @dlm: the display list manager * @header: display list header * @extension: extended display list header. NULL for normal lists * @dma: DMA address for the header * @body0: first display list body * @bodies: list of extra display list bodies * @pre_cmd: pre command to be issued through extended dl header * @post_cmd: post command to be issued through extended dl header * @allocated: flag to detect double list release * @has_chain: if true, indicates that there's a partition chain * @chain: entry in the display list partition chain * @flags: display list flags, a combination of VSP1_DL_FRAME_END_*
*/ struct vsp1_dl_list { struct list_head list; struct vsp1_dl_manager *dlm;
/** * struct vsp1_dl_manager - Display List manager * @index: index of the related WPF * @singleshot: execute the display list in single-shot mode * @vsp1: the VSP1 device * @lock: protects the free, active, queued, and pending lists * @free: array of all free display lists * @active: list currently being processed (loaded) by hardware * @queued: list queued to the hardware (written to the DL registers) * @pending: list waiting to be queued to the hardware * @pool: body pool for the display list bodies * @cmdpool: commands pool for extended display list * @list_count: number of allocated display lists
*/ struct vsp1_dl_manager { unsignedint index; bool singleshot; struct vsp1_device *vsp1;
/* ----------------------------------------------------------------------------- * Display List Body Management
*/
/** * vsp1_dl_body_pool_create - Create a pool of bodies from a single allocation * @vsp1: The VSP1 device * @num_bodies: The number of bodies to allocate * @num_entries: The maximum number of entries that a body can contain * @extra_size: Extra allocation provided for the bodies * * Allocate a pool of display list bodies each with enough memory to contain the * requested number of entries plus the @extra_size. * * Return a pointer to a pool on success or NULL if memory can't be allocated.
*/ struct vsp1_dl_body_pool *
vsp1_dl_body_pool_create(struct vsp1_device *vsp1, unsignedint num_bodies, unsignedint num_entries, size_t extra_size)
{ struct vsp1_dl_body_pool *pool;
size_t dlb_size; unsignedint i;
pool = kzalloc(sizeof(*pool), GFP_KERNEL); if (!pool) return NULL;
pool->vsp1 = vsp1;
/* * TODO: 'extra_size' is only used by vsp1_dlm_create(), to allocate * extra memory for the display list header. We need only one header per * display list, not per display list body, thus this allocation is * extraneous and should be reworked in the future.
*/
dlb_size = num_entries * sizeof(struct vsp1_dl_entry) + extra_size;
pool->size = dlb_size * num_bodies;
for (i = 0; i < num_bodies; ++i) { struct vsp1_dl_body *dlb = &pool->bodies[i];
dlb->pool = pool;
dlb->max_entries = num_entries;
dlb->dma = pool->dma + i * dlb_size;
dlb->entries = pool->mem + i * dlb_size;
list_add_tail(&dlb->free, &pool->free);
}
return pool;
}
/** * vsp1_dl_body_pool_destroy - Release a body pool * @pool: The body pool * * Release all components of a pool allocation.
*/ void vsp1_dl_body_pool_destroy(struct vsp1_dl_body_pool *pool)
{ if (!pool) return;
if (pool->mem)
dma_free_wc(pool->vsp1->bus_master, pool->size, pool->mem,
pool->dma);
kfree(pool->bodies);
kfree(pool);
}
/** * vsp1_dl_body_get - Obtain a body from a pool * @pool: The body pool * * Obtain a body from the pool without blocking. * * Returns a display list body or NULL if there are none available.
*/ struct vsp1_dl_body *vsp1_dl_body_get(struct vsp1_dl_body_pool *pool)
{ struct vsp1_dl_body *dlb = NULL; unsignedlong flags;
/** * vsp1_dl_body_put - Return a body back to its pool * @dlb: The display list body * * Return a body back to the pool, and reset the num_entries to clear the list.
*/ void vsp1_dl_body_put(struct vsp1_dl_body *dlb)
{ unsignedlong flags;
/** * vsp1_dl_body_write - Write a register to a display list body * @dlb: The body * @reg: The register address * @data: The register value * * Write the given register and value to the display list body. The maximum * number of entries that can be written in a body is specified when the body is * allocated by vsp1_dl_body_alloc().
*/ void vsp1_dl_body_write(struct vsp1_dl_body *dlb, u32 reg, u32 data)
{ if (WARN_ONCE(dlb->num_entries >= dlb->max_entries, "DLB size exceeded (max %u)", dlb->max_entries)) return;
/** * vsp1_dl_cmd_pool_create - Create a pool of commands from a single allocation * @vsp1: The VSP1 device * @type: The command pool type * @num_cmds: The number of commands to allocate * * Allocate a pool of commands each with enough memory to contain the private * data of each command. The allocation sizes are dependent upon the command * type. * * Return a pointer to the pool on success or NULL if memory can't be allocated.
*/ staticstruct vsp1_dl_cmd_pool *
vsp1_dl_cmd_pool_create(struct vsp1_device *vsp1, enum vsp1_extcmd_type type, unsignedint num_cmds)
{ struct vsp1_dl_cmd_pool *pool; unsignedint i;
size_t cmd_size;
pool = kzalloc(sizeof(*pool), GFP_KERNEL); if (!pool) return NULL;
/* * TODO: Auto-disp can utilise more than one extended body * command per cmd.
*/
cmd->num_cmds = 1;
cmd->cmds = pool->mem + cmd_offset;
cmd->cmd_dma = pool->dma + cmd_offset;
/** * vsp1_dl_list_get - Get a free display list * @dlm: The display list manager * * Get a display list from the pool of free lists and return it. * * This function must be called without the display list manager lock held.
*/ struct vsp1_dl_list *vsp1_dl_list_get(struct vsp1_dl_manager *dlm)
{ struct vsp1_dl_list *dl = NULL; unsignedlong flags;
lockdep_assert_not_held(&dlm->lock);
spin_lock_irqsave(&dlm->lock, flags);
if (!list_empty(&dlm->free)) {
dl = list_first_entry(&dlm->free, struct vsp1_dl_list, list);
list_del(&dl->list);
/* * The display list chain must be initialised to ensure every * display list can assert list_empty() if it is not in a chain.
*/
INIT_LIST_HEAD(&dl->chain);
dl->allocated = true;
}
spin_unlock_irqrestore(&dlm->lock, flags);
return dl;
}
/* This function must be called with the display list manager lock held.*/ staticvoid __vsp1_dl_list_put(struct vsp1_dl_list *dl)
{ struct vsp1_dl_list *dl_next;
if (!dl) return;
lockdep_assert_held(&dl->dlm->lock);
/* * Release any linked display-lists which were chained for a single * hardware operation.
*/ if (dl->has_chain) {
list_for_each_entry(dl_next, &dl->chain, chain)
__vsp1_dl_list_put(dl_next);
}
/* * body0 is reused as as an optimisation as presently every display list * has at least one body, thus we reinitialise the entries list.
*/
dl->body0->num_entries = 0;
/* * Return the display list to the 'free' pool. If the list had already * been returned be loud about it.
*/
WARN_ON_ONCE(!dl->allocated);
dl->allocated = false;
list_add_tail(&dl->list, &dl->dlm->free);
}
/** * vsp1_dl_list_put - Release a display list * @dl: The display list * * Release the display list and return it to the pool of free lists. * * Passing a NULL pointer to this function is safe, in that case no operation * will be performed.
*/ void vsp1_dl_list_put(struct vsp1_dl_list *dl)
{ unsignedlong flags;
/** * vsp1_dl_list_get_body0 - Obtain the default body for the display list * @dl: The display list * * Obtain a pointer to the internal display list body allowing this to be passed * directly to configure operations.
*/ struct vsp1_dl_body *vsp1_dl_list_get_body0(struct vsp1_dl_list *dl)
{ return dl->body0;
}
/** * vsp1_dl_list_add_body - Add a body to the display list * @dl: The display list * @dlb: The body * * Add a display list body to a display list. Registers contained in bodies are * processed after registers contained in the main display list, in the order in * which bodies are added. * * Adding a body to a display list passes ownership of the body to the list. The * caller retains its reference to the body when adding it to the display list, * but is not allowed to add new entries to the body. * * The reference must be explicitly released by a call to vsp1_dl_body_put() * when the body isn't needed anymore.
*/ int vsp1_dl_list_add_body(struct vsp1_dl_list *dl, struct vsp1_dl_body *dlb)
{
refcount_inc(&dlb->refcnt);
list_add_tail(&dlb->list, &dl->bodies);
return 0;
}
/** * vsp1_dl_list_add_chain - Add a display list to a chain * @head: The head display list * @dl: The new display list * * Add a display list to an existing display list chain. The chained lists * will be automatically processed by the hardware without intervention from * the CPU. A display list end interrupt will only complete after the last * display list in the chain has completed processing. * * Adding a display list to a chain passes ownership of the display list to * the head display list item. The chain is released when the head dl item is * put back with __vsp1_dl_list_put().
*/ int vsp1_dl_list_add_chain(struct vsp1_dl_list *head, struct vsp1_dl_list *dl)
{
head->has_chain = true;
list_add_tail(&dl->chain, &head->chain); return 0;
}
/* * Fill the header with the display list bodies addresses and sizes. The * address of the first body has already been filled when the display * list was allocated.
*/
/* * Enable the interrupt for the end of each frame. In continuous mode * chained lists are used with one list per frame, so enable the * interrupt for each list. In singleshot mode chained lists are used * to partition a single frame, so enable the interrupt for the last * list only.
*/ if (!dlm->singleshot || is_last)
dl->header->flags |= VSP1_DLH_INT_ENABLE;
/* * In continuous mode enable auto-start for all lists, as the VSP must * loop on the same list until a new one is queued. In singleshot mode * enable auto-start for all lists but the last to chain processing of * partitions without software intervention.
*/ if (!dlm->singleshot || !is_last)
dl->header->flags |= VSP1_DLH_AUTO_START;
if (!is_last) { /* * If this is not the last display list in the chain, queue the * next item for automatic processing by the hardware.
*/ struct vsp1_dl_list *next = list_next_entry(dl, chain);
dl->header->next_header = next->dma;
} elseif (!dlm->singleshot) { /* * if the display list manager works in continuous mode, the VSP * should loop over the display list continuously until * instructed to do otherwise.
*/
dl->header->next_header = dl->dma;
}
/* * Check whether the VSP1 has taken the update. The hardware indicates * this by clearing the UPDHDR bit in the CMD register.
*/ return !!(vsp1_read(vsp1, VI6_CMD(dlm->index)) & VI6_CMD_UPDHDR);
}
/* * Program the display list header address. If the hardware is idle * (single-shot mode or first frame in continuous mode) it will then be * started independently. If the hardware is operating, the * VI6_DL_HDR_REF_ADDR register will be updated with the display list * address.
*/
vsp1_write(vsp1, VI6_DL_HDR_ADDR(dlm->index), dl->dma);
}
/* * If a previous display list has been queued to the hardware but not * processed yet, the VSP can start processing it at any time. In that * case we can't replace the queued list by the new one, as we could * race with the hardware. We thus mark the update as pending, it will * be queued up to the hardware by the frame end interrupt handler. * * If a display list is already pending we simply drop it as the new * display list is assumed to contain a more recent configuration. It is * an error if the already pending list has the * VSP1_DL_FRAME_END_INTERNAL flag set, as there is then a process * waiting for that list to complete. This shouldn't happen as the * waiting process should perform proper locking, but warn just in * case.
*/ if (vsp1_dl_list_hw_update_pending(dlm)) {
WARN_ON(dlm->pending &&
(dlm->pending->flags & VSP1_DL_FRAME_END_INTERNAL));
__vsp1_dl_list_put(dlm->pending);
dlm->pending = dl; return;
}
/* * Pass the new display list to the hardware and mark it as queued. It * will become active when the hardware starts processing it.
*/
vsp1_dl_list_hw_enqueue(dl);
/* * When working in single-shot mode, the caller guarantees that the * hardware is idle at this point. Just commit the head display list * to hardware. Chained lists will be started automatically.
*/
vsp1_dl_list_hw_enqueue(dl);
if (dlm->singleshot)
vsp1_dl_list_commit_singleshot(dl); else
vsp1_dl_list_commit_continuous(dl);
spin_unlock_irqrestore(&dlm->lock, flags);
}
/* ----------------------------------------------------------------------------- * Display List Manager
*/
/** * vsp1_dlm_irq_frame_end - Display list handler for the frame end interrupt * @dlm: the display list manager * * Return a set of flags that indicates display list completion status. * * The VSP1_DL_FRAME_END_COMPLETED flag indicates that the previous display list * has completed at frame end. If the flag is not returned display list * completion has been delayed by one frame because the display list commit * raced with the frame end interrupt. The function always returns with the flag * set in single-shot mode as display list processing is then not continuous and * races never occur. * * The following flags are only supported for continuous mode. * * The VSP1_DL_FRAME_END_INTERNAL flag indicates that the display list that just * became active had been queued with the internal notification flag. * * The VSP1_DL_FRAME_END_WRITEBACK flag indicates that the previously active * display list had been queued with the writeback flag.
*/ unsignedint vsp1_dlm_irq_frame_end(struct vsp1_dl_manager *dlm)
{ struct vsp1_device *vsp1 = dlm->vsp1;
u32 status = vsp1_read(vsp1, VI6_STATUS); unsignedint flags = 0;
spin_lock(&dlm->lock);
/* * The mem-to-mem pipelines work in single-shot mode. No new display * list can be queued, we don't have to do anything.
*/ if (dlm->singleshot) {
__vsp1_dl_list_put(dlm->active);
dlm->active = NULL;
flags |= VSP1_DL_FRAME_END_COMPLETED; goto done;
}
/* * If the commit operation raced with the interrupt and occurred after * the frame end event but before interrupt processing, the hardware * hasn't taken the update into account yet. We have to skip one frame * and retry.
*/ if (vsp1_dl_list_hw_update_pending(dlm)) goto done;
/* * Progressive streams report only TOP fields. If we have a BOTTOM * field, we are interlaced, and expect the frame to complete on the * next frame end interrupt.
*/ if (status & VI6_STATUS_FLD_STD(dlm->index)) goto done;
/* * If the active display list has the writeback flag set, the frame * completion marks the end of the writeback capture. Return the * VSP1_DL_FRAME_END_WRITEBACK flag and reset the display list's * writeback flag.
*/ if (dlm->active && (dlm->active->flags & VSP1_DL_FRAME_END_WRITEBACK)) {
flags |= VSP1_DL_FRAME_END_WRITEBACK;
dlm->active->flags &= ~VSP1_DL_FRAME_END_WRITEBACK;
}
/* * The device starts processing the queued display list right after the * frame end interrupt. The display list thus becomes active.
*/ if (dlm->queued) { if (dlm->queued->flags & VSP1_DL_FRAME_END_INTERNAL)
flags |= VSP1_DL_FRAME_END_INTERNAL;
dlm->queued->flags &= ~VSP1_DL_FRAME_END_INTERNAL;
/* * Now that the VSP has started processing the queued display list, we * can queue the pending display list to the hardware if one has been * prepared.
*/ if (dlm->pending) {
vsp1_dl_list_hw_enqueue(dlm->pending);
dlm->queued = dlm->pending;
dlm->pending = NULL;
}
/* * Initialize the display list body and allocate DMA memory for the body * and the header. Both are allocated together to avoid memory * fragmentation, with the header located right after the body in * memory. An extra body is allocated on top of the prealloc to account * for the cached body used by the vsp1_pipeline object.
*/
header_size = vsp1_feature(vsp1, VSP1_HAS_EXT_DL) ? sizeof(struct vsp1_dl_header_extended) : sizeof(struct vsp1_dl_header);
Die Informationen auf dieser Webseite wurden
nach bestem Wissen sorgfältig zusammengestellt. Es wird jedoch weder Vollständigkeit, noch Richtigkeit,
noch Qualität der bereit gestellten Informationen zugesichert.
Bemerkung:
Die farbliche Syntaxdarstellung und die Messung sind noch experimentell.