/* * Defined as bits to be able to use them as a mask on Rx. * Also used as internal return values on Tx.
*/ enum {
LIBETH_XDP_PASS = 0U,
LIBETH_XDP_DROP = BIT(0),
LIBETH_XDP_ABORTED = BIT(1),
LIBETH_XDP_TX = BIT(2),
LIBETH_XDP_REDIRECT = BIT(3),
};
/* * &xdp_buff_xsk is the largest structure &libeth_xdp_buff gets casted to, * pick maximum pointer-compatible alignment.
*/ #define __LIBETH_XDP_BUFF_ALIGN \
(IS_ALIGNED(sizeof(struct xdp_buff_xsk), 16) ? 16 : \
IS_ALIGNED(sizeof(struct xdp_buff_xsk), 8) ? 8 : \ sizeof(long))
/** * struct libeth_xdp_buff - libeth extension over &xdp_buff * @base: main &xdp_buff * @data: shortcut for @base.data * @desc: RQ descriptor containing metadata for this buffer * @priv: driver-private scratchspace * * The main reason for this is to have a pointer to the descriptor to be able * to quickly get frame metadata from xdpmo and driver buff-to-xdp callbacks * (as well as bigger alignment). * Pointer/layout-compatible with &xdp_buff and &xdp_buff_xsk.
*/ struct libeth_xdp_buff { union { struct xdp_buff base; void *data;
};
/** * __LIBETH_XDP_ONSTACK_BUFF - declare a &libeth_xdp_buff on the stack * @name: name of the variable to declare * @...: sizeof() of the driver-private data
*/ #define __LIBETH_XDP_ONSTACK_BUFF(name, ...) \
___LIBETH_XDP_ONSTACK_BUFF(name, ##__VA_ARGS__) /** * LIBETH_XDP_ONSTACK_BUFF - declare a &libeth_xdp_buff on the stack * @name: name of the variable to declare * @...: type or variable name of the driver-private data
*/ #define LIBETH_XDP_ONSTACK_BUFF(name, ...) \
__LIBETH_XDP_ONSTACK_BUFF(name, __libeth_xdp_priv_sz(__VA_ARGS__))
/** * libeth_xdpsq_num - calculate optimal number of XDPSQs for this device + sys * @rxq: current number of active Rx queues * @txq: current number of active Tx queues * @max: maximum number of Tx queues * * Each RQ must have its own XDPSQ for XSk pairs, each CPU must have own XDPSQ * for lockless sending (``XDP_TX``, .ndo_xdp_xmit()). Cap the maximum of these * two with the number of SQs the device can have (minus used ones). * * Return: number of XDP Tx queues the device needs to use.
*/ staticinline u32 libeth_xdpsq_num(u32 rxq, u32 txq, u32 max)
{ return min(max(nr_cpu_ids, rxq), max - txq);
}
/** * libeth_xdpsq_shared - whether XDPSQs can be shared between several CPUs * @num: number of active XDPSQs * * Return: true if there's no 1:1 XDPSQ/CPU association, false otherwise.
*/ staticinlinebool libeth_xdpsq_shared(u32 num)
{ return num < nr_cpu_ids;
}
/** * libeth_xdpsq_id - get XDPSQ index corresponding to this CPU * @num: number of active XDPSQs * * Helper for libeth_xdp routines, do not use in drivers directly. * * Return: XDPSQ index needs to be used on this CPU.
*/ staticinline u32 libeth_xdpsq_id(u32 num)
{
u32 ret = raw_smp_processor_id();
if (static_branch_unlikely(&libeth_xdpsq_share) &&
libeth_xdpsq_shared(num))
ret %= num;
/** * libeth_xdpsq_get - initialize &libeth_xdpsq_lock * @lock: lock to initialize * @dev: netdev which this lock belongs to * @share: whether XDPSQs can be shared * * Tracks the current XDPSQ association and enables the static lock * if needed.
*/ staticinlinevoid libeth_xdpsq_get(struct libeth_xdpsq_lock *lock, conststruct net_device *dev, bool share)
{ if (unlikely(share))
__libeth_xdpsq_get(lock, dev);
}
/** * libeth_xdpsq_put - deinitialize &libeth_xdpsq_lock * @lock: lock to deinitialize * @dev: netdev which this lock belongs to * * Tracks the current XDPSQ association and disables the static lock * if needed.
*/ staticinlinevoid libeth_xdpsq_put(struct libeth_xdpsq_lock *lock, conststruct net_device *dev)
{ if (static_branch_unlikely(&libeth_xdpsq_share) && lock->share)
__libeth_xdpsq_put(lock, dev);
}
/** * libeth_xdpsq_lock - grab &libeth_xdpsq_lock if needed * @lock: lock to take * * Touches the underlying spinlock only if the static key is enabled * and the queue itself is marked as shareable.
*/ staticinlinevoid libeth_xdpsq_lock(struct libeth_xdpsq_lock *lock)
{ if (static_branch_unlikely(&libeth_xdpsq_share) && lock->share)
__libeth_xdpsq_lock(lock);
}
/** * libeth_xdpsq_unlock - free &libeth_xdpsq_lock if needed * @lock: lock to free * * Touches the underlying spinlock only if the static key is enabled * and the queue itself is marked as shareable.
*/ staticinlinevoid libeth_xdpsq_unlock(struct libeth_xdpsq_lock *lock)
{ if (static_branch_unlikely(&libeth_xdpsq_share) && lock->share)
__libeth_xdpsq_unlock(lock);
}
/** * libeth_xdpsq_deinit_timer - deinitialize &libeth_xdpsq_timer * @timer: timer to deinitialize * * Flush and disable the underlying workqueue.
*/ staticinlinevoid libeth_xdpsq_deinit_timer(struct libeth_xdpsq_timer *timer)
{
cancel_delayed_work_sync(&timer->dwork);
}
/** * libeth_xdpsq_queue_timer - run &libeth_xdpsq_timer * @timer: timer to queue * * Should be called after the queue was filled and the transmission was run * to complete the pending buffers if no further sending will be done in a * second (-> lazy cleaning won't happen). * If the timer was already run, it will be requeued back to one second * timeout again.
*/ staticinlinevoid libeth_xdpsq_queue_timer(struct libeth_xdpsq_timer *timer)
{
mod_delayed_work_on(raw_smp_processor_id(), system_bh_highpri_wq,
&timer->dwork, HZ);
}
/** * libeth_xdpsq_run_timer - wrapper to run a queue clean-up on a timer event * @work: workqueue belonging to the corresponding timer * @poll: driver-specific completion queue poll function * * Run the polling function on the locked queue and requeue the timer if * there's more work to do. * Designed to be used via LIBETH_XDP_DEFINE_TIMER() below.
*/ static __always_inline void
libeth_xdpsq_run_timer(struct work_struct *work,
u32 (*poll)(void *xdpsq, u32 budget))
{ struct libeth_xdpsq_timer *timer = container_of(work, typeof(*timer),
dwork.work);
libeth_xdpsq_lock(timer->lock);
if (poll(timer->xdpsq, U32_MAX))
libeth_xdpsq_queue_timer(timer);
libeth_xdpsq_unlock(timer->lock);
}
/* Common Tx bits */
/** * enum - libeth_xdp internal Tx flags * @LIBETH_XDP_TX_BULK: one bulk size at which it will be flushed to the queue * @LIBETH_XDP_TX_BATCH: batch size for which the queue fill loop is unrolled * @LIBETH_XDP_TX_DROP: indicates the send function must drop frames not sent * @LIBETH_XDP_TX_NDO: whether the send function is called from .ndo_xdp_xmit() * @LIBETH_XDP_TX_XSK: whether the function is called for ``XDP_TX`` for XSk
*/ enum {
LIBETH_XDP_TX_BULK = DEV_MAP_BULK_SIZE,
LIBETH_XDP_TX_BATCH = 8,
/** * enum - &libeth_xdp_tx_frame and &libeth_xdp_tx_desc flags * @LIBETH_XDP_TX_LEN: only for ``XDP_TX``, [15:0] of ::len_fl is actual length * @LIBETH_XDP_TX_CSUM: for XSk xmit, enable checksum offload * @LIBETH_XDP_TX_XSKMD: for XSk xmit, mask of the metadata bits * @LIBETH_XDP_TX_FIRST: indicates the frag is the first one of the frame * @LIBETH_XDP_TX_LAST: whether the frag is the last one of the frame * @LIBETH_XDP_TX_MULTI: whether the frame contains several frags * @LIBETH_XDP_TX_FLAGS: only for ``XDP_TX``, [31:16] of ::len_fl is flags
*/ enum {
LIBETH_XDP_TX_LEN = GENMASK(15, 0),
/** * struct libeth_xdp_tx_frame - represents one XDP Tx element * @data: frame start pointer for ``XDP_TX`` * @len_fl: ``XDP_TX``, combined flags [31:16] and len [15:0] field for speed * @soff: ``XDP_TX``, offset from @data to the start of &skb_shared_info * @frag: one (non-head) frag for ``XDP_TX`` * @xdpf: &xdp_frame for the head frag for .ndo_xdp_xmit() * @dma: DMA address of the non-head frag for .ndo_xdp_xmit() * @xsk: ``XDP_TX`` for XSk, XDP buffer for any frag * @len: frag length for XSk ``XDP_TX`` and .ndo_xdp_xmit() * @flags: Tx flags for the above * @opts: combined @len + @flags for the above for speed * @desc: XSk xmit descriptor for direct casting
*/ struct libeth_xdp_tx_frame { union { /* ``XDP_TX`` */ struct { void *data;
u32 len_fl;
u32 soff;
};
/** * struct libeth_xdp_tx_bulk - XDP Tx frame bulk for bulk sending * @prog: corresponding active XDP program, %NULL for .ndo_xdp_xmit() * @dev: &net_device which the frames are transmitted on * @xdpsq: shortcut to the corresponding driver-specific XDPSQ structure * @act_mask: Rx only, mask of all the XDP prog verdicts for that NAPI session * @count: current number of frames in @bulk * @bulk: array of queued frames for bulk Tx * * All XDP Tx operations except XSk xmit queue each frame to the bulk first * and flush it when @count reaches the array end. Bulk is always placed on * the stack for performance. One bulk element contains all the data necessary * for sending a frame and then freeing it on completion. * For XSk xmit, Tx descriptor array from &xsk_buff_pool is casted directly * to &libeth_xdp_tx_frame as they are compatible and the bulk structure is * not used.
*/ struct libeth_xdp_tx_bulk { conststruct bpf_prog *prog; struct net_device *dev; void *xdpsq;
/** * LIBETH_XDP_ONSTACK_BULK - declare &libeth_xdp_tx_bulk on the stack * @bq: name of the variable to declare * * Helper to declare a bulk on the stack with a compiler hint that it should * not be initialized automatically (with `CONFIG_INIT_STACK_ALL_*`) for * performance reasons.
*/ #define LIBETH_XDP_ONSTACK_BULK(bq) \ struct libeth_xdp_tx_bulk bq __uninitialized
/** * struct libeth_xdpsq - abstraction for an XDPSQ * @pool: XSk buffer pool for XSk ``XDP_TX`` and xmit * @sqes: array of Tx buffers from the actual queue struct * @descs: opaque pointer to the HW descriptor array * @ntu: pointer to the next free descriptor index * @count: number of descriptors on that queue * @pending: pointer to the number of sent-not-completed descs on that queue * @xdp_tx: pointer to the above, but only for non-XSk-xmit frames * @lock: corresponding XDPSQ lock * * Abstraction for driver-independent implementation of Tx. Placed on the stack * and filled by the driver before the transmission, so that the generic * functions can access and modify driver-specific resources.
*/ struct libeth_xdpsq { struct xsk_buff_pool *pool; struct libeth_sqe *sqes; void *descs;
/** * struct libeth_xdp_tx_desc - abstraction for an XDP Tx descriptor * @addr: DMA address of the frame * @len: length of the frame * @flags: XDP Tx flags * @opts: combined @len + @flags for speed * * Filled by the generic functions and then passed to driver-specific functions * to fill a HW Tx descriptor, always placed on the [function] stack.
*/ struct libeth_xdp_tx_desc {
dma_addr_t addr; union { struct {
u32 len;
u32 flags;
};
aligned_u64 opts;
};
} __aligned_largest;
/** * libeth_xdp_ptr_to_priv - convert pointer to a libeth_xdp u64 priv * @ptr: pointer to convert * * The main sending function passes private data as the largest scalar, u64. * Use this helper when you want to pass a pointer there.
*/ #define libeth_xdp_ptr_to_priv(ptr) ({ \
typecheck_pointer(ptr); \
((u64)(uintptr_t)(ptr)); \
}) /** * libeth_xdp_priv_to_ptr - convert libeth_xdp u64 priv to a pointer * @priv: private data to convert * * The main sending function passes private data as the largest scalar, u64. * Use this helper when your callback takes this u64 and you want to convert * it back to a pointer.
*/ #define libeth_xdp_priv_to_ptr(priv) ({ \
static_assert(__same_type(priv, u64)); \
((constvoid *)(uintptr_t)(priv)); \
})
/* * On 64-bit systems, assigning one u64 is faster than two u32s. When ::len * occupies lowest 32 bits (LE), whole ::opts can be assigned directly instead.
*/ #ifdef __LITTLE_ENDIAN #define __LIBETH_WORD_ACCESS 1 #endif #ifdef __LIBETH_WORD_ACCESS #define __libeth_xdp_tx_len(flen, ...) \
.opts = ((flen) | FIELD_PREP(GENMASK_ULL(63, 32), (__VA_ARGS__ + 0))) #else #define __libeth_xdp_tx_len(flen, ...) \
.len = (flen), .flags = (__VA_ARGS__ + 0) #endif
/** * libeth_xdp_tx_xmit_bulk - main XDP Tx function * @bulk: array of frames to send * @xdpsq: pointer to the driver-specific XDPSQ struct * @n: number of frames to send * @unroll: whether to unroll the queue filling loop for speed * @priv: driver-specific private data * @prep: callback for cleaning the queue and filling abstract &libeth_xdpsq * @fill: internal callback for filling &libeth_sqe and &libeth_xdp_tx_desc * @xmit: callback for filling a HW descriptor with the frame info * * Internal abstraction for placing @n XDP Tx frames on the HW XDPSQ. Used for * all types of frames: ``XDP_TX``, .ndo_xdp_xmit(), XSk ``XDP_TX``, and XSk * xmit. * @prep must lock the queue as this function releases it at the end. @unroll * greatly increases the object code size, but also greatly increases XSk xmit * performance; for other types of frames, it's not enabled. * The compilers inline all those onstack abstractions to direct data accesses. * * Return: number of frames actually placed on the queue, <= @n. The function * can't fail, but can send less frames if there's no enough free descriptors * available. The actual free space is returned by @prep from the driver.
*/ static __always_inline u32
libeth_xdp_tx_xmit_bulk(conststruct libeth_xdp_tx_frame *bulk, void *xdpsq,
u32 n, bool unroll, u64 priv,
u32 (*prep)(void *xdpsq, struct libeth_xdpsq *sq), struct libeth_xdp_tx_desc
(*fill)(struct libeth_xdp_tx_frame frm, u32 i, conststruct libeth_xdpsq *sq, u64 priv), void (*xmit)(struct libeth_xdp_tx_desc desc, u32 i, conststruct libeth_xdpsq *sq, u64 priv))
{ struct libeth_xdpsq sq __uninitialized;
u32 this, batched, off = 0;
u32 ntu, i = 0;
n = min(n, prep(xdpsq, &sq)); if (unlikely(!n)) goto unlock;
ntu = *sq.ntu;
this = sq.count - ntu; if (likely(this > n)) this = n;
again: if (!unroll) goto linear;
batched = ALIGN_DOWN(this, LIBETH_XDP_TX_BATCH);
for ( ; i < off + batched; i += LIBETH_XDP_TX_BATCH) {
u32 base = ntu + i - off;
unrolled_count(LIBETH_XDP_TX_BATCH) for (u32 j = 0; j < LIBETH_XDP_TX_BATCH; j++)
xmit(fill(bulk[i + j], base + j, &sq, priv),
base + j, &sq, priv);
}
if (batched < this) {
linear: for ( ; i < off + this; i++)
xmit(fill(bulk[i], ntu + i - off, &sq, priv),
ntu + i - off, &sq, priv);
}
ntu += this; if (likely(ntu < sq.count)) goto out;
/** * libeth_xdp_tx_queue_head - internal helper for queueing one ``XDP_TX`` head * @bq: XDP Tx bulk to queue the head frag to * @xdp: XDP buffer with the head to queue * * Return: false if it's the only frag of the frame, true if it's an S/G frame.
*/ staticinlinebool libeth_xdp_tx_queue_head(struct libeth_xdp_tx_bulk *bq, conststruct libeth_xdp_buff *xdp)
{ conststruct xdp_buff *base = &xdp->base;
/** * libeth_xdp_tx_fill_stats - fill &libeth_sqe with ``XDP_TX`` frame stats * @sqe: SQ element to fill * @desc: libeth_xdp Tx descriptor * @sinfo: &skb_shared_info for this frame * * Internal helper for filling an SQE with the frame stats, do not use in * drivers. Fills the number of frags and bytes for this frame.
*/ #define libeth_xdp_tx_fill_stats(sqe, desc, sinfo) \
__libeth_xdp_tx_fill_stats(sqe, desc, sinfo, __UNIQUE_ID(sqe_), \
__UNIQUE_ID(desc_), __UNIQUE_ID(sinfo_))
/** * libeth_xdp_tx_fill_buf - internal helper to fill one ``XDP_TX`` &libeth_sqe * @frm: XDP Tx frame from the bulk * @i: index on the HW queue * @sq: XDPSQ abstraction for the queue * @priv: private data * * Return: XDP Tx descriptor with the synced DMA and other info to pass to * the driver callback.
*/ staticinlinestruct libeth_xdp_tx_desc
libeth_xdp_tx_fill_buf(struct libeth_xdp_tx_frame frm, u32 i, conststruct libeth_xdpsq *sq, u64 priv)
{ struct libeth_xdp_tx_desc desc; struct skb_shared_info *sinfo;
skb_frag_t *frag = &frm.frag; struct libeth_sqe *sqe;
netmem_ref netmem;
/** * libeth_xdp_tx_flush_bulk - wrapper to define flush of one ``XDP_TX`` bulk * @bq: bulk to flush * @flags: Tx flags, see above * @prep: driver callback to prepare the queue * @xmit: driver callback to fill a HW descriptor * * Use via LIBETH_XDP_DEFINE_FLUSH_TX() to define an ``XDP_TX`` driver * callback.
*/ #define libeth_xdp_tx_flush_bulk(bq, flags, prep, xmit) \
__libeth_xdp_tx_flush_bulk(bq, flags, prep, libeth_xdp_tx_fill_buf, \
xmit)
/* .ndo_xdp_xmit() implementation */
/** * libeth_xdp_xmit_init_bulk - internal helper to initialize bulk for XDP xmit * @bq: bulk to initialize * @dev: target &net_device * @xdpsqs: array of driver-specific XDPSQ structs * @num: number of active XDPSQs (the above array length)
*/ #define libeth_xdp_xmit_init_bulk(bq, dev, xdpsqs, num) \
__libeth_xdp_xmit_init_bulk(bq, dev, (xdpsqs)[libeth_xdpsq_id(num)])
/** * libeth_xdp_xmit_frame_dma - internal helper to access DMA of an &xdp_frame * @xf: pointer to the XDP frame * * There's no place in &libeth_xdp_tx_frame to store DMA address for an * &xdp_frame head. The headroom is used then, the address is placed right * after the frame struct, naturally aligned. * * Return: pointer to the DMA address to use.
*/ #define libeth_xdp_xmit_frame_dma(xf) \
_Generic((xf), \ conststruct xdp_frame *: \
(const dma_addr_t *)__libeth_xdp_xmit_frame_dma(xf), \ struct xdp_frame *: \
(dma_addr_t *)__libeth_xdp_xmit_frame_dma(xf) \
)
if (!IS_ENABLED(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) &&
__alignof(*xdpf) < sizeof(dma_addr_t))
addr = PTR_ALIGN(addr, sizeof(dma_addr_t));
return addr;
}
/** * libeth_xdp_xmit_queue_head - internal helper for queueing one XDP xmit head * @bq: XDP Tx bulk to queue the head frag to * @xdpf: XDP frame with the head to queue * @dev: device to perform DMA mapping * * Return: ``LIBETH_XDP_DROP`` on DMA mapping error, * ``LIBETH_XDP_PASS`` if it's the only frag in the frame, * ``LIBETH_XDP_TX`` if it's an S/G frame.
*/ staticinline u32 libeth_xdp_xmit_queue_head(struct libeth_xdp_tx_bulk *bq, struct xdp_frame *xdpf, struct device *dev)
{
dma_addr_t dma;
/** * libeth_xdp_xmit_queue_bulk - internal helper for queueing one XDP xmit frame * @bq: XDP Tx bulk to queue the frame to * @xdpf: XDP frame to queue * @flush_bulk: driver callback to flush the bulk to the HW queue * * Return: ``LIBETH_XDP_TX`` on success, * ``LIBETH_XDP_DROP`` if the frame should be dropped by the stack, * ``LIBETH_XDP_ABORTED`` if the frame will be dropped by libeth_xdp.
*/ static __always_inline u32
libeth_xdp_xmit_queue_bulk(struct libeth_xdp_tx_bulk *bq, struct xdp_frame *xdpf, bool (*flush_bulk)(struct libeth_xdp_tx_bulk *bq,
u32 flags))
{
u32 head, nr_frags, i, ret = LIBETH_XDP_TX; struct device *dev = bq->dev->dev.parent; conststruct skb_shared_info *sinfo;
if (unlikely(bq->count == LIBETH_XDP_TX_BULK) &&
unlikely(!flush_bulk(bq, LIBETH_XDP_TX_NDO))) return LIBETH_XDP_DROP;
head = libeth_xdp_xmit_queue_head(bq, xdpf, dev); if (head == LIBETH_XDP_PASS) goto out; elseif (head == LIBETH_XDP_DROP) return LIBETH_XDP_DROP;
/** * libeth_xdp_xmit_fill_buf - internal helper to fill one XDP xmit &libeth_sqe * @frm: XDP Tx frame from the bulk * @i: index on the HW queue * @sq: XDPSQ abstraction for the queue * @priv: private data * * Return: XDP Tx descriptor with the mapped DMA and other info to pass to * the driver callback.
*/ staticinlinestruct libeth_xdp_tx_desc
libeth_xdp_xmit_fill_buf(struct libeth_xdp_tx_frame frm, u32 i, conststruct libeth_xdpsq *sq, u64 priv)
{ struct libeth_xdp_tx_desc desc; struct libeth_sqe *sqe; struct xdp_frame *xdpf;
/** * libeth_xdp_xmit_flush_bulk - wrapper to define flush of one XDP xmit bulk * @bq: bulk to flush * @flags: Tx flags, see __libeth_xdp_tx_flush_bulk() * @prep: driver callback to prepare the queue * @xmit: driver callback to fill a HW descriptor * * Use via LIBETH_XDP_DEFINE_FLUSH_XMIT() to define an XDP xmit driver * callback.
*/ #define libeth_xdp_xmit_flush_bulk(bq, flags, prep, xmit) \
__libeth_xdp_tx_flush_bulk(bq, (flags) | LIBETH_XDP_TX_NDO, prep, \
libeth_xdp_xmit_fill_buf, xmit)
/** * __libeth_xdp_xmit_do_bulk - internal function to implement .ndo_xdp_xmit() * @bq: XDP Tx bulk to queue frames to * @frames: XDP frames passed by the stack * @n: number of frames * @flags: flags passed by the stack * @flush_bulk: driver callback to flush an XDP xmit bulk * @finalize: driver callback to finalize sending XDP Tx frames on the queue * * Perform common checks, map the frags and queue them to the bulk, then flush * the bulk to the XDPSQ. If requested by the stack, finalize the queue. * * Return: number of frames send or -errno on error.
*/ static __always_inline int
__libeth_xdp_xmit_do_bulk(struct libeth_xdp_tx_bulk *bq, struct xdp_frame **frames, u32 n, u32 flags, bool (*flush_bulk)(struct libeth_xdp_tx_bulk *bq,
u32 flags), void (*finalize)(void *xdpsq, bool sent, bool flush))
{
u32 nxmit = 0;
if (unlikely(flags & ~XDP_XMIT_FLAGS_MASK)) return -EINVAL;
for (u32 i = 0; likely(i < n); i++) {
u32 ret;
ret = libeth_xdp_xmit_queue_bulk(bq, frames[i], flush_bulk); if (unlikely(ret != LIBETH_XDP_TX)) {
nxmit += ret == LIBETH_XDP_ABORTED; break;
}
nxmit++;
}
if (bq->count) {
flush_bulk(bq, LIBETH_XDP_TX_NDO); if (unlikely(bq->count))
nxmit -= libeth_xdp_xmit_return_bulk(bq->bulk,
bq->count,
bq->dev);
}
/** * libeth_xdp_xmit_do_bulk - implement full .ndo_xdp_xmit() in driver * @dev: target &net_device * @n: number of frames to send * @fr: XDP frames to send * @f: flags passed by the stack * @xqs: array of XDPSQs driver structs * @nqs: number of active XDPSQs, the above array length * @fl: driver callback to flush an XDP xmit bulk * @fin: driver cabback to finalize the queue * * If the driver has active XDPSQs, perform common checks and send the frames. * Finalize the queue, if requested. * * Return: number of frames sent or -errno on error.
*/ #define libeth_xdp_xmit_do_bulk(dev, n, fr, f, xqs, nqs, fl, fin) \
_libeth_xdp_xmit_do_bulk(dev, n, fr, f, xqs, nqs, fl, fin, \
__UNIQUE_ID(bq_), __UNIQUE_ID(ret_), \
__UNIQUE_ID(nqs_))
#define _libeth_xdp_xmit_do_bulk(d, n, fr, f, xqs, nqs, fl, fin, ub, ur, un) \
({ \
u32 un = (nqs); \ int ur; \
\ if (likely(un)) { \
LIBETH_XDP_ONSTACK_BULK(ub); \
\
libeth_xdp_xmit_init_bulk(&ub, d, xqs, un); \
ur = __libeth_xdp_xmit_do_bulk(&ub, fr, n, f, fl, fin); \
} else { \
ur = -ENXIO; \
} \
\
ur; \
})
/* Rx polling path */
/** * libeth_xdp_tx_init_bulk - initialize an XDP Tx bulk for Rx NAPI poll * @bq: bulk to initialize * @prog: RCU pointer to the XDP program (can be %NULL) * @dev: target &net_device * @xdpsqs: array of driver XDPSQ structs * @num: number of active XDPSQs, the above array length * * Should be called on an onstack XDP Tx bulk before the NAPI polling loop. * Initializes all the needed fields to run libeth_xdp functions. If @num == 0, * assumes XDP is not enabled. * Do not use for XSk, it has its own optimized helper.
*/ #define libeth_xdp_tx_init_bulk(bq, prog, dev, xdpsqs, num) \
__libeth_xdp_tx_init_bulk(bq, prog, dev, xdpsqs, num, false, \
__UNIQUE_ID(bq_), __UNIQUE_ID(nqs_))
/** * libeth_xdp_init_buff - initialize a &libeth_xdp_buff for Rx NAPI poll * @dst: onstack buffer to initialize * @src: XDP buffer stash placed on the queue * @rxq: registered &xdp_rxq_info corresponding to this queue * * Should be called before the main NAPI polling loop. Loads the content of * the previously saved stash or initializes the buffer from scratch. * Do not use for XSk.
*/ staticinlinevoid
libeth_xdp_init_buff(struct libeth_xdp_buff *dst, conststruct libeth_xdp_buff_stash *src, struct xdp_rxq_info *rxq)
{ if (likely(!src->data))
dst->data = NULL; else
libeth_xdp_load_stash(dst, src);
dst->base.rxq = rxq;
}
/** * libeth_xdp_save_buff - save a partially built buffer on a queue * @dst: XDP buffer stash placed on the queue * @src: onstack buffer to save * * Should be called after the main NAPI polling loop. If the loop exited before * the buffer was finished, saves its content on the queue, so that it can be * completed during the next poll. Otherwise, clears the stash.
*/ staticinlinevoid libeth_xdp_save_buff(struct libeth_xdp_buff_stash *dst, conststruct libeth_xdp_buff *src)
{ if (likely(!src->data))
dst->data = NULL; else
libeth_xdp_save_stash(dst, src);
}
/** * libeth_xdp_return_stash - free an XDP buffer stash from a queue * @stash: stash to free * * If the queue is about to be destroyed, but it still has an incompleted * buffer stash, this helper should be called to free it.
*/ staticinlinevoid libeth_xdp_return_stash(struct libeth_xdp_buff_stash *stash)
{ if (stash->data)
__libeth_xdp_return_stash(stash);
}
/** * libeth_xdp_prepare_buff - fill &libeth_xdp_buff with head FQE data * @xdp: XDP buffer to attach the head to * @fqe: FQE containing the head buffer * @len: buffer len passed from HW * * Internal, use libeth_xdp_process_buff() instead. Initializes XDP buffer * head with the Rx buffer data: data pointer, length, headroom, and * truesize/tailroom. Zeroes the flags. * Uses faster single u64 write instead of per-field access.
*/ staticinlinevoid libeth_xdp_prepare_buff(struct libeth_xdp_buff *xdp, conststruct libeth_fqe *fqe,
u32 len)
{ conststruct page *page = __netmem_to_page(fqe->netmem);
/** * libeth_xdp_process_buff - attach Rx buffer to &libeth_xdp_buff * @xdp: XDP buffer to attach the Rx buffer to * @fqe: Rx buffer to process * @len: received data length from the descriptor * * If the XDP buffer is empty, attaches the Rx buffer as head and initializes * the required fields. Otherwise, attaches the buffer as a frag. * Already performs DMA sync-for-CPU and frame start prefetch * (for head buffers only). * * Return: true on success, false if the descriptor must be skipped (empty or * no space for a new frag).
*/ staticinlinebool libeth_xdp_process_buff(struct libeth_xdp_buff *xdp, conststruct libeth_fqe *fqe,
u32 len)
{ if (!libeth_rx_sync_for_cpu(fqe, len)) returnfalse;
if (xdp->data) return libeth_xdp_buff_add_frag(xdp, fqe, len);
libeth_xdp_prepare_buff(xdp, fqe, len);
prefetch(xdp->data);
returntrue;
}
/** * libeth_xdp_buff_stats_frags - update onstack RQ stats with XDP frags info * @ss: onstack stats to update * @xdp: buffer to account * * Internal helper used by __libeth_xdp_run_pass(), do not call directly. * Adds buffer's frags count and total len to the onstack stats.
*/ staticinlinevoid
libeth_xdp_buff_stats_frags(struct libeth_rq_napi_stats *ss, conststruct libeth_xdp_buff *xdp)
{ conststruct skb_shared_info *sinfo;
/** * __libeth_xdp_run_prog - run XDP program on an XDP buffer * @xdp: XDP buffer to run the prog on * @bq: buffer bulk for ``XDP_TX`` queueing * * Internal inline abstraction to run XDP program. Handles ``XDP_DROP`` * and ``XDP_REDIRECT`` only, the rest is processed levels up. * Reports an XDP prog exception on errors. * * Return: libeth_xdp prog verdict depending on the prog's verdict.
*/ static __always_inline u32
__libeth_xdp_run_prog(struct libeth_xdp_buff *xdp, conststruct libeth_xdp_tx_bulk *bq)
{ enum xdp_action act;
/** * __libeth_xdp_run_flush - run XDP program and handle ``XDP_TX`` verdict * @xdp: XDP buffer to run the prog on * @bq: buffer bulk for ``XDP_TX`` queueing * @run: internal callback for running XDP program * @queue: internal callback for queuing ``XDP_TX`` frame * @flush_bulk: driver callback for flushing a bulk * * Internal inline abstraction to run XDP program and additionally handle * ``XDP_TX`` verdict. Used by both XDP and XSk, hence @run and @queue. * Do not use directly. * * Return: libeth_xdp prog verdict depending on the prog's verdict.
*/ static __always_inline u32
__libeth_xdp_run_flush(struct libeth_xdp_buff *xdp, struct libeth_xdp_tx_bulk *bq,
u32 (*run)(struct libeth_xdp_buff *xdp, conststruct libeth_xdp_tx_bulk *bq), bool (*queue)(struct libeth_xdp_tx_bulk *bq, struct libeth_xdp_buff *xdp, bool (*flush_bulk)
(struct libeth_xdp_tx_bulk *bq,
u32 flags)), bool (*flush_bulk)(struct libeth_xdp_tx_bulk *bq,
u32 flags))
{
u32 act;
/** * libeth_xdp_run_prog - run XDP program (non-XSk path) and handle all verdicts * @xdp: XDP buffer to process * @bq: XDP Tx bulk to queue ``XDP_TX`` buffers * @fl: driver ``XDP_TX`` bulk flush callback * * Run the attached XDP program and handle all possible verdicts. XSk has its * own version. * Prefer using it via LIBETH_XDP_DEFINE_RUN{,_PASS,_PROG}(). * * Return: true if the buffer should be passed up the stack, false if the poll * should go to the next buffer.
*/ #define libeth_xdp_run_prog(xdp, bq, fl) \
(__libeth_xdp_run_flush(xdp, bq, __libeth_xdp_run_prog, \
libeth_xdp_tx_queue_bulk, \
fl) == LIBETH_XDP_PASS)
/** * __libeth_xdp_run_pass - helper to run XDP program and handle the result * @xdp: XDP buffer to process * @bq: XDP Tx bulk to queue ``XDP_TX`` frames * @napi: NAPI to build an skb and pass it up the stack * @rs: onstack libeth RQ stats * @md: metadata that should be filled to the XDP buffer * @prep: callback for filling the metadata * @run: driver wrapper to run XDP program * @populate: driver callback to populate an skb with the HW descriptor data * * Inline abstraction that does the following (non-XSk path): * 1) adds frame size and frag number (if needed) to the onstack stats; * 2) fills the descriptor metadata to the onstack &libeth_xdp_buff * 3) runs XDP program if present; * 4) handles all possible verdicts; * 5) on ``XDP_PASS`, builds an skb from the buffer; * 6) populates it with the descriptor metadata; * 7) passes it up the stack. * * In most cases, number 2 means just writing the pointer to the HW descriptor * to the XDP buffer. If so, please use LIBETH_XDP_DEFINE_RUN{,_PASS}() * wrappers to build a driver function.
*/ static __always_inline void
__libeth_xdp_run_pass(struct libeth_xdp_buff *xdp, struct libeth_xdp_tx_bulk *bq, struct napi_struct *napi, struct libeth_rq_napi_stats *rs, constvoid *md, void (*prep)(struct libeth_xdp_buff *xdp, constvoid *md), bool (*run)(struct libeth_xdp_buff *xdp, struct libeth_xdp_tx_bulk *bq), bool (*populate)(struct sk_buff *skb, conststruct libeth_xdp_buff *xdp, struct libeth_rq_napi_stats *rs))
{ struct sk_buff *skb;
/** * libeth_xdp_run_pass - helper to run XDP program and handle the result * @xdp: XDP buffer to process * @bq: XDP Tx bulk to queue ``XDP_TX`` frames * @napi: NAPI to build an skb and pass it up the stack * @ss: onstack libeth RQ stats * @desc: pointer to the HW descriptor for that frame * @run: driver wrapper to run XDP program * @populate: driver callback to populate an skb with the HW descriptor data * * Wrapper around the underscored version when "fill the descriptor metadata" * means just writing the pointer to the HW descriptor as @xdp->desc.
*/ #define libeth_xdp_run_pass(xdp, bq, napi, ss, desc, run, populate) \
__libeth_xdp_run_pass(xdp, bq, napi, ss, desc, libeth_xdp_prep_desc, \
run, populate)
/** * libeth_xdp_finalize_rx - finalize XDPSQ after a NAPI polling loop (non-XSk) * @bq: ``XDP_TX`` frame bulk * @flush: driver callback to flush the bulk * @finalize: driver callback to start sending the frames and run the timer * * Flush the bulk if there are frames left to send, kick the queue and flush * the XDP maps.
*/ #define libeth_xdp_finalize_rx(bq, flush, finalize) \
__libeth_xdp_finalize_rx(bq, 0, flush, finalize)
/* * Helpers to reduce boilerplate code in drivers. * * Typical driver Rx flow would be (excl. bulk and buff init, frag attach): * * LIBETH_XDP_DEFINE_START(); * LIBETH_XDP_DEFINE_FLUSH_TX(static driver_xdp_flush_tx, driver_xdp_tx_prep, * driver_xdp_xmit); * LIBETH_XDP_DEFINE_RUN(static driver_xdp_run, driver_xdp_run_prog, * driver_xdp_flush_tx, driver_populate_skb); * LIBETH_XDP_DEFINE_FINALIZE(static driver_xdp_finalize_rx, * driver_xdp_flush_tx, driver_xdp_finalize_sq); * LIBETH_XDP_DEFINE_END(); * * This will build a set of 4 static functions. The compiler is free to decide * whether to inline them. * Then, in the NAPI polling function: * * while (packets < budget) { * // ... * driver_xdp_run(xdp, &bq, napi, &rs, desc); * } * driver_xdp_finalize_rx(&bq);
*/
#define LIBETH_XDP_DEFINE_START() \
__diag_push(); \
__diag_ignore(GCC, 8, "-Wold-style-declaration", \ "Allow specifying \'static\' after the return type")
/** * LIBETH_XDP_DEFINE_TIMER - define a driver XDPSQ cleanup timer callback * @name: name of the function to define * @poll: Tx polling/completion function
*/ #define LIBETH_XDP_DEFINE_TIMER(name, poll) \ void name(struct work_struct *work) \
{ \
libeth_xdpsq_run_timer(work, poll); \
}
/** * LIBETH_XDP_DEFINE_FLUSH_TX - define a driver ``XDP_TX`` bulk flush function * @name: name of the function to define * @prep: driver callback to clean an XDPSQ * @xmit: driver callback to write a HW Tx descriptor
*/ #define LIBETH_XDP_DEFINE_FLUSH_TX(name, prep, xmit) \
__LIBETH_XDP_DEFINE_FLUSH_TX(name, prep, xmit, xdp)
/** * LIBETH_XDP_DEFINE_FLUSH_XMIT - define a driver XDP xmit bulk flush function * @name: name of the function to define * @prep: driver callback to clean an XDPSQ * @xmit: driver callback to write a HW Tx descriptor
*/ #define LIBETH_XDP_DEFINE_FLUSH_XMIT(name, prep, xmit) \ bool name(struct libeth_xdp_tx_bulk *bq, u32 flags) \
{ \ return libeth_xdp_xmit_flush_bulk(bq, flags, prep, xmit); \
}
/** * LIBETH_XDP_DEFINE_RUN_PROG - define a driver XDP program run function * @name: name of the function to define * @flush: driver callback to flush an ``XDP_TX`` bulk
*/ #define LIBETH_XDP_DEFINE_RUN_PROG(name, flush) \ bool __LIBETH_XDP_DEFINE_RUN_PROG(name, flush, xdp)
/** * LIBETH_XDP_DEFINE_RUN_PASS - define a driver buffer process + pass function * @name: name of the function to define * @run: driver callback to run XDP program (above) * @populate: driver callback to fill an skb with HW descriptor info
*/ #define LIBETH_XDP_DEFINE_RUN_PASS(name, run, populate) \ void __LIBETH_XDP_DEFINE_RUN_PASS(name, run, populate, xdp)
/** * LIBETH_XDP_DEFINE_RUN - define a driver buffer process, run + pass function * @name: name of the function to define * @run: name of the XDP prog run function to define * @flush: driver callback to flush an ``XDP_TX`` bulk * @populate: driver callback to fill an skb with HW descriptor info
*/ #define LIBETH_XDP_DEFINE_RUN(name, run, flush, populate) \
__LIBETH_XDP_DEFINE_RUN(name, run, flush, populate, XDP)
/** * LIBETH_XDP_DEFINE_FINALIZE - define a driver Rx NAPI poll finalize function * @name: name of the function to define * @flush: driver callback to flush an ``XDP_TX`` bulk * @finalize: driver callback to finalize an XDPSQ and run the timer
*/ #define LIBETH_XDP_DEFINE_FINALIZE(name, flush, finalize) \
__LIBETH_XDP_DEFINE_FINALIZE(name, flush, finalize, xdp)
/** * libeth_xdp_buff_to_rq - get RQ pointer from an XDP buffer pointer * @xdp: &libeth_xdp_buff corresponding to the queue * @type: typeof() of the driver Rx queue structure * @member: name of &xdp_rxq_info inside @type * * Often times, pointer to the RQ is needed when reading/filling metadata from * HW descriptors. The helper can be used to quickly jump from an XDP buffer * to the queue corresponding to its &xdp_rxq_info without introducing * additional fields (&libeth_xdp_buff is precisely 1 cacheline long on x64).
*/ #define libeth_xdp_buff_to_rq(xdp, type, member) \
container_of_const((xdp)->base.rxq, type, member)
/** * libeth_xdpmo_rx_hash - convert &libeth_rx_pt to an XDP RSS hash metadata * @hash: pointer to the variable to write the hash to * @rss_type: pointer to the variable to write the hash type to * @val: hash value from the HW descriptor * @pt: libeth parsed packet type * * Handle zeroed/non-available hash and convert libeth parsed packet type to * the corresponding XDP RSS hash type. To be called at the end of * xdp_metadata_ops idpf_xdpmo::xmo_rx_hash() implementation. * Note that if the driver doesn't use a constant packet type lookup table but * generates it at runtime, it must call libeth_rx_pt_gen_hash_type(pt) to * generate XDP RSS hash type for each packet type. * * Return: 0 on success, -ENODATA when the hash is not available.
*/ staticinlineint libeth_xdpmo_rx_hash(u32 *hash, enum xdp_rss_hash_type *rss_type,
u32 val, struct libeth_rx_pt pt)
{ if (unlikely(!val)) return -ENODATA;
/** * __libeth_xdp_complete_tx - complete sent XDPSQE * @sqe: SQ element / Tx buffer to complete * @cp: Tx polling/completion params * @bulk: internal callback to bulk-free ``XDP_TX`` buffers * @xsk: internal callback to free XSk ``XDP_TX`` buffers * * Use the non-underscored version in drivers instead. This one is shared * internally with libeth_tx_complete_any(). * Complete an XDPSQE of any type of XDP frame. This includes DMA unmapping * when needed, buffer freeing, stats update, and SQE invalidation.
*/ static __always_inline void
__libeth_xdp_complete_tx(struct libeth_sqe *sqe, struct libeth_cq_pp *cp,
typeof(libeth_xdp_return_buff_bulk) bulk,
typeof(libeth_xsk_buff_free_slow) xsk)
{ enum libeth_sqe_type type = sqe->type;
switch (type) { case LIBETH_SQE_EMPTY: return; case LIBETH_SQE_XDP_XMIT: case LIBETH_SQE_XDP_XMIT_FRAG:
dma_unmap_page(cp->dev, dma_unmap_addr(sqe, dma),
dma_unmap_len(sqe, len), DMA_TO_DEVICE); break; default: break;
}
switch (type) { case LIBETH_SQE_XDP_TX:
bulk(sqe->sinfo, cp->bq, sqe->nr_frags != 1); break; case LIBETH_SQE_XDP_XMIT:
xdp_return_frame_bulk(sqe->xdpf, cp->bq); break; case LIBETH_SQE_XSK_TX: case LIBETH_SQE_XSK_TX_FRAG:
xsk(sqe->xsk); break; default: break;
}
switch (type) { case LIBETH_SQE_XDP_TX: case LIBETH_SQE_XDP_XMIT: case LIBETH_SQE_XSK_TX:
cp->xdp_tx -= sqe->nr_frags;
/** * libeth_xdp_set_features - set XDP features for netdev * @dev: &net_device to configure * @...: optional params, see __libeth_xdp_set_features() * * Set all the features libeth_xdp supports, including .ndo_xdp_xmit(). That * said, it should be used only when XDPSQs are always available regardless * of whether an XDP prog is attached to @dev.
*/ #define libeth_xdp_set_features(dev, ...) \
CONCATENATE(__libeth_xdp_feat, \
COUNT_ARGS(__VA_ARGS__))(dev, ##__VA_ARGS__)
/** * libeth_xdp_set_features_noredir - enable all libeth_xdp features w/o redir * @dev: target &net_device * @...: optional params, see __libeth_xdp_set_features() * * Enable everything except the .ndo_xdp_xmit() feature, use when XDPSQs are * not available right after netdev registration.
*/ #define libeth_xdp_set_features_noredir(dev, ...) \
__libeth_xdp_set_features_noredir(dev, __UNIQUE_ID(dev_), \ ##__VA_ARGS__)
Die Informationen auf dieser Webseite wurden
nach bestem Wissen sorgfältig zusammengestellt. Es wird jedoch weder Vollständigkeit, noch Richtigkeit,
noch Qualität der bereit gestellten Informationen zugesichert.
Bemerkung:
Die farbliche Syntaxdarstellung und die Messung sind noch experimentell.