/** * idpf_ctlq_setup_regs - initialize control queue registers * @cq: pointer to the specific control queue * @q_create_info: structs containing info for each queue to be initialized
*/ staticvoid idpf_ctlq_setup_regs(struct idpf_ctlq_info *cq, struct idpf_ctlq_create_info *q_create_info)
{ /* set control queue registers in our local struct */
cq->reg.head = q_create_info->reg.head;
cq->reg.tail = q_create_info->reg.tail;
cq->reg.len = q_create_info->reg.len;
cq->reg.bah = q_create_info->reg.bah;
cq->reg.bal = q_create_info->reg.bal;
cq->reg.len_mask = q_create_info->reg.len_mask;
cq->reg.len_ena_mask = q_create_info->reg.len_ena_mask;
cq->reg.head_mask = q_create_info->reg.head_mask;
}
/** * idpf_ctlq_init_regs - Initialize control queue registers * @hw: pointer to hw struct * @cq: pointer to the specific Control queue * @is_rxq: true if receive control queue, false otherwise * * Initialize registers. The caller is expected to have already initialized the * descriptor ring memory and buffer memory
*/ staticvoid idpf_ctlq_init_regs(struct idpf_hw *hw, struct idpf_ctlq_info *cq, bool is_rxq)
{ /* Update tail to post pre-allocated buffers for rx queues */ if (is_rxq)
idpf_mbx_wr32(hw, cq->reg.tail, (u32)(cq->ring_size - 1));
/* For non-Mailbox control queues only TAIL need to be set */ if (cq->q_id != -1) return;
/* Clear Head for both send or receive */
idpf_mbx_wr32(hw, cq->reg.head, 0);
/* set starting point */
idpf_mbx_wr32(hw, cq->reg.bal, lower_32_bits(cq->desc_ring.pa));
idpf_mbx_wr32(hw, cq->reg.bah, upper_32_bits(cq->desc_ring.pa));
idpf_mbx_wr32(hw, cq->reg.len, (cq->ring_size | cq->reg.len_ena_mask));
}
/** * idpf_ctlq_init_rxq_bufs - populate receive queue descriptors with buf * @cq: pointer to the specific Control queue * * Record the address of the receive queue DMA buffers in the descriptors. * The buffers must have been previously allocated.
*/ staticvoid idpf_ctlq_init_rxq_bufs(struct idpf_ctlq_info *cq)
{ int i;
for (i = 0; i < cq->ring_size; i++) { struct idpf_ctlq_desc *desc = IDPF_CTLQ_DESC(cq, i); struct idpf_dma_mem *bi = cq->bi.rx_buff[i];
/* No buffer to post to descriptor, continue */ if (!bi) continue;
/** * idpf_ctlq_shutdown - shutdown the CQ * @hw: pointer to hw struct * @cq: pointer to the specific Control queue * * The main shutdown routine for any controq queue
*/ staticvoid idpf_ctlq_shutdown(struct idpf_hw *hw, struct idpf_ctlq_info *cq)
{
spin_lock(&cq->cq_lock);
/* free ring buffers and the ring itself */
idpf_ctlq_dealloc_ring_res(hw, cq);
/* Set ring_size to 0 to indicate uninitialized queue */
cq->ring_size = 0;
spin_unlock(&cq->cq_lock);
}
/** * idpf_ctlq_add - add one control queue * @hw: pointer to hardware struct * @qinfo: info for queue to be created * @cq_out: (output) double pointer to control queue to be created * * Allocate and initialize a control queue and add it to the control queue list. * The cq parameter will be allocated/initialized and passed back to the caller * if no errors occur. * * Note: idpf_ctlq_init must be called prior to any calls to idpf_ctlq_add
*/ int idpf_ctlq_add(struct idpf_hw *hw, struct idpf_ctlq_create_info *qinfo, struct idpf_ctlq_info **cq_out)
{ struct idpf_ctlq_info *cq; bool is_rxq = false; int err;
cq = kzalloc(sizeof(*cq), GFP_KERNEL); if (!cq) return -ENOMEM;
if (is_rxq) {
idpf_ctlq_init_rxq_bufs(cq);
} else { /* Allocate the array of msg pointers for TX queues */
cq->bi.tx_msg = kcalloc(qinfo->len, sizeof(struct idpf_ctlq_msg *),
GFP_KERNEL); if (!cq->bi.tx_msg) {
err = -ENOMEM; goto init_dealloc_q_mem;
}
}
idpf_ctlq_setup_regs(cq, qinfo);
idpf_ctlq_init_regs(hw, cq, is_rxq);
spin_lock_init(&cq->cq_lock);
list_add(&cq->cq_list, &hw->cq_list_head);
*cq_out = cq;
return 0;
init_dealloc_q_mem: /* free ring buffers and the ring itself */
idpf_ctlq_dealloc_ring_res(hw, cq);
init_free_q:
kfree(cq);
return err;
}
/** * idpf_ctlq_remove - deallocate and remove specified control queue * @hw: pointer to hardware struct * @cq: pointer to control queue to be removed
*/ void idpf_ctlq_remove(struct idpf_hw *hw, struct idpf_ctlq_info *cq)
{
list_del(&cq->cq_list);
idpf_ctlq_shutdown(hw, cq);
kfree(cq);
}
/** * idpf_ctlq_init - main initialization routine for all control queues * @hw: pointer to hardware struct * @num_q: number of queues to initialize * @q_info: array of structs containing info for each queue to be initialized * * This initializes any number and any type of control queues. This is an all * or nothing routine; if one fails, all previously allocated queues will be * destroyed. This must be called prior to using the individual add/remove * APIs.
*/ int idpf_ctlq_init(struct idpf_hw *hw, u8 num_q, struct idpf_ctlq_create_info *q_info)
{ struct idpf_ctlq_info *cq, *tmp; int err; int i;
INIT_LIST_HEAD(&hw->cq_list_head);
for (i = 0; i < num_q; i++) { struct idpf_ctlq_create_info *qinfo = q_info + i;
err = idpf_ctlq_add(hw, qinfo, &cq); if (err) goto init_destroy_qs;
}
/** * idpf_ctlq_send - send command to Control Queue (CTQ) * @hw: pointer to hw struct * @cq: handle to control queue struct to send on * @num_q_msg: number of messages to send on control queue * @q_msg: pointer to array of queue messages to be sent * * The caller is expected to allocate DMAable buffers and pass them to the * send routine via the q_msg struct / control queue specific data struct. * The control queue will hold a reference to each send message until * the completion for that message has been cleaned.
*/ int idpf_ctlq_send(struct idpf_hw *hw, struct idpf_ctlq_info *cq,
u16 num_q_msg, struct idpf_ctlq_msg q_msg[])
{ struct idpf_ctlq_desc *desc; int num_desc_avail; int err = 0; int i;
spin_lock(&cq->cq_lock);
/* Ensure there are enough descriptors to send all messages */
num_desc_avail = IDPF_CTLQ_DESC_UNUSED(cq); if (num_desc_avail == 0 || num_desc_avail < num_q_msg) {
err = -ENOSPC; goto err_unlock;
}
for (i = 0; i < num_q_msg; i++) { struct idpf_ctlq_msg *msg = &q_msg[i];
/* Update the address values in the desc with the pa * value for respective buffer
*/
desc->params.indirect.addr_high =
cpu_to_le32(upper_32_bits(buff->pa));
desc->params.indirect.addr_low =
cpu_to_le32(lower_32_bits(buff->pa));
/* Store buffer info */
cq->bi.tx_msg[cq->next_to_use] = msg;
(cq->next_to_use)++; if (cq->next_to_use == cq->ring_size)
cq->next_to_use = 0;
}
/* Force memory write to complete before letting hardware * know that there are new descriptors to fetch.
*/
dma_wmb();
idpf_mbx_wr32(hw, cq->reg.tail, cq->next_to_use);
err_unlock:
spin_unlock(&cq->cq_lock);
return err;
}
/** * idpf_ctlq_clean_sq - reclaim send descriptors on HW write back for the * requested queue * @cq: pointer to the specific Control queue * @clean_count: (input|output) number of descriptors to clean as input, and * number of descriptors actually cleaned as output * @msg_status: (output) pointer to msg pointer array to be populated; needs * to be allocated by caller * * Returns an array of message pointers associated with the cleaned * descriptors. The pointers are to the original ctlq_msgs sent on the cleaned * descriptors. The status will be returned for each; any messages that failed * to send will have a non-zero status. The caller is expected to free original * ctlq_msgs and free or reuse the DMA buffers.
*/ int idpf_ctlq_clean_sq(struct idpf_ctlq_info *cq, u16 *clean_count, struct idpf_ctlq_msg *msg_status[])
{ struct idpf_ctlq_desc *desc;
u16 i, num_to_clean;
u16 ntc, desc_err;
if (*clean_count == 0) return 0; if (*clean_count > cq->ring_size) return -EBADR;
spin_lock(&cq->cq_lock);
ntc = cq->next_to_clean;
num_to_clean = *clean_count;
for (i = 0; i < num_to_clean; i++) { /* Fetch next descriptor and check if marked as done */
desc = IDPF_CTLQ_DESC(cq, ntc); if (!(le16_to_cpu(desc->flags) & IDPF_CTLQ_FLAG_DD)) break;
/* Ensure no other fields are read until DD flag is checked */
dma_rmb();
/* Zero out any stale data */
memset(desc, 0, sizeof(*desc));
ntc++; if (ntc == cq->ring_size)
ntc = 0;
}
cq->next_to_clean = ntc;
spin_unlock(&cq->cq_lock);
/* Return number of descriptors actually cleaned */
*clean_count = i;
return 0;
}
/** * idpf_ctlq_post_rx_buffs - post buffers to descriptor ring * @hw: pointer to hw struct * @cq: pointer to control queue handle * @buff_count: (input|output) input is number of buffers caller is trying to * return; output is number of buffers that were not posted * @buffs: array of pointers to dma mem structs to be given to hardware * * Caller uses this function to return DMA buffers to the descriptor ring after * consuming them; buff_count will be the number of buffers. * * Note: this function needs to be called after a receive call even * if there are no DMA buffers to be returned, i.e. buff_count = 0, * buffs = NULL to support direct commands
*/ int idpf_ctlq_post_rx_buffs(struct idpf_hw *hw, struct idpf_ctlq_info *cq,
u16 *buff_count, struct idpf_dma_mem **buffs)
{ struct idpf_ctlq_desc *desc;
u16 ntp = cq->next_to_post; bool buffs_avail = false;
u16 tbp = ntp + 1; int i = 0;
if (*buff_count > cq->ring_size) return -EBADR;
if (*buff_count > 0)
buffs_avail = true;
spin_lock(&cq->cq_lock);
if (tbp >= cq->ring_size)
tbp = 0;
if (tbp == cq->next_to_clean) /* Nothing to do */ goto post_buffs_out;
/* Post buffers for as many as provided or up until the last one used */ while (ntp != cq->next_to_clean) {
desc = IDPF_CTLQ_DESC(cq, ntp);
if (cq->bi.rx_buff[ntp]) goto fill_desc; if (!buffs_avail) { /* If the caller hasn't given us any buffers or * there are none left, search the ring itself * for an available buffer to move to this * entry starting at the next entry in the ring
*/
tbp = ntp + 1;
/* Wrap ring if necessary */ if (tbp >= cq->ring_size)
tbp = 0;
while (tbp != cq->next_to_clean) { if (cq->bi.rx_buff[tbp]) {
cq->bi.rx_buff[ntp] =
cq->bi.rx_buff[tbp];
cq->bi.rx_buff[tbp] = NULL;
/* Found a buffer, no need to * search anymore
*/ break;
}
/* Wrap ring if necessary */
tbp++; if (tbp >= cq->ring_size)
tbp = 0;
}
if (tbp == cq->next_to_clean) goto post_buffs_out;
} else { /* Give back pointer to DMA buffer */
cq->bi.rx_buff[ntp] = buffs[i];
i++;
/* Post buffers to descriptor */
desc->datalen = cpu_to_le16(cq->bi.rx_buff[ntp]->size);
desc->params.indirect.addr_high =
cpu_to_le32(upper_32_bits(cq->bi.rx_buff[ntp]->pa));
desc->params.indirect.addr_low =
cpu_to_le32(lower_32_bits(cq->bi.rx_buff[ntp]->pa));
ntp++; if (ntp == cq->ring_size)
ntp = 0;
}
post_buffs_out: /* Only update tail if buffers were actually posted */ if (cq->next_to_post != ntp) { if (ntp) /* Update next_to_post to ntp - 1 since current ntp * will not have a buffer
*/
cq->next_to_post = ntp - 1; else /* Wrap to end of end ring since current ntp is 0 */
cq->next_to_post = cq->ring_size - 1;
/* return the number of buffers that were not posted */
*buff_count = *buff_count - i;
return 0;
}
/** * idpf_ctlq_recv - receive control queue message call back * @cq: pointer to control queue handle to receive on * @num_q_msg: (input|output) input number of messages that should be received; * output number of messages actually received * @q_msg: (output) array of received control queue messages on this q; * needs to be pre-allocated by caller for as many messages as requested * * Called by interrupt handler or polling mechanism. Caller is expected * to free buffers
*/ int idpf_ctlq_recv(struct idpf_ctlq_info *cq, u16 *num_q_msg, struct idpf_ctlq_msg *q_msg)
{
u16 num_to_clean, ntc, flags; struct idpf_ctlq_desc *desc; int err = 0;
u16 i;
/* take the lock before we start messing with the ring */
spin_lock(&cq->cq_lock);
ntc = cq->next_to_clean;
num_to_clean = *num_q_msg;
for (i = 0; i < num_to_clean; i++) { /* Fetch next descriptor and check if marked as done */
desc = IDPF_CTLQ_DESC(cq, ntc);
flags = le16_to_cpu(desc->flags);
if (!(flags & IDPF_CTLQ_FLAG_DD)) break;
/* Ensure no other fields are read until DD flag is checked */
dma_rmb();
if (desc->datalen) {
memcpy(q_msg[i].ctx.indirect.context,
&desc->params.indirect, IDPF_INDIRECT_CTX_SIZE);
/* Assign pointer to dma buffer to ctlq_msg array * to be given to upper layer
*/
q_msg[i].ctx.indirect.payload = cq->bi.rx_buff[ntc];
/* Zero out pointer to DMA buffer info; * will be repopulated by post buffers API
*/
cq->bi.rx_buff[ntc] = NULL;
} else {
memcpy(q_msg[i].ctx.direct, desc->params.raw,
IDPF_DIRECT_CTX_SIZE);
}
/* Zero out stale data in descriptor */
memset(desc, 0, sizeof(struct idpf_ctlq_desc));
ntc++; if (ntc == cq->ring_size)
ntc = 0;
}
cq->next_to_clean = ntc;
spin_unlock(&cq->cq_lock);
*num_q_msg = i; if (*num_q_msg == 0)
err = -ENOMSG;
return err;
}
Messung V0.5
¤ Dauer der Verarbeitung: 0.10 Sekunden
(vorverarbeitet)
¤
Die Informationen auf dieser Webseite wurden
nach bestem Wissen sorgfältig zusammengestellt. Es wird jedoch weder Vollständigkeit, noch Richtigkeit,
noch Qualität der bereit gestellten Informationen zugesichert.
Bemerkung:
Die farbliche Syntaxdarstellung und die Messung sind noch experimentell.