mirror of
git://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
synced 2025-08-05 16:54:27 +00:00

To eliminate the use of struct page in page pool, the page pool users should use netmem descriptor and APIs instead. Make xdp access ->pp through netmem_desc instead of page. Signed-off-by: Byungchul Park <byungchul@sk.com> Link: https://patch.msgid.link/20250721021835.63939-13-byungchul@sk.com Signed-off-by: Jakub Kicinski <kuba@kernel.org>
1879 lines
56 KiB
C
1879 lines
56 KiB
C
/* SPDX-License-Identifier: GPL-2.0-only */
|
|
/* Copyright (C) 2025 Intel Corporation */
|
|
|
|
#ifndef __LIBETH_XDP_H
|
|
#define __LIBETH_XDP_H
|
|
|
|
#include <linux/bpf_trace.h>
|
|
#include <linux/unroll.h>
|
|
|
|
#include <net/libeth/rx.h>
|
|
#include <net/libeth/tx.h>
|
|
#include <net/xsk_buff_pool.h>
|
|
|
|
/*
|
|
* Defined as bits to be able to use them as a mask on Rx.
|
|
* Also used as internal return values on Tx.
|
|
*/
|
|
enum {
|
|
LIBETH_XDP_PASS = 0U,
|
|
LIBETH_XDP_DROP = BIT(0),
|
|
LIBETH_XDP_ABORTED = BIT(1),
|
|
LIBETH_XDP_TX = BIT(2),
|
|
LIBETH_XDP_REDIRECT = BIT(3),
|
|
};
|
|
|
|
/*
|
|
* &xdp_buff_xsk is the largest structure &libeth_xdp_buff gets casted to,
|
|
* pick maximum pointer-compatible alignment.
|
|
*/
|
|
#define __LIBETH_XDP_BUFF_ALIGN \
|
|
(IS_ALIGNED(sizeof(struct xdp_buff_xsk), 16) ? 16 : \
|
|
IS_ALIGNED(sizeof(struct xdp_buff_xsk), 8) ? 8 : \
|
|
sizeof(long))
|
|
|
|
/**
|
|
* struct libeth_xdp_buff - libeth extension over &xdp_buff
|
|
* @base: main &xdp_buff
|
|
* @data: shortcut for @base.data
|
|
* @desc: RQ descriptor containing metadata for this buffer
|
|
* @priv: driver-private scratchspace
|
|
*
|
|
* The main reason for this is to have a pointer to the descriptor to be able
|
|
* to quickly get frame metadata from xdpmo and driver buff-to-xdp callbacks
|
|
* (as well as bigger alignment).
|
|
* Pointer/layout-compatible with &xdp_buff and &xdp_buff_xsk.
|
|
*/
|
|
struct libeth_xdp_buff {
|
|
union {
|
|
struct xdp_buff base;
|
|
void *data;
|
|
};
|
|
|
|
const void *desc;
|
|
unsigned long priv[]
|
|
__aligned(__LIBETH_XDP_BUFF_ALIGN);
|
|
} __aligned(__LIBETH_XDP_BUFF_ALIGN);
|
|
static_assert(offsetof(struct libeth_xdp_buff, data) ==
|
|
offsetof(struct xdp_buff_xsk, xdp.data));
|
|
static_assert(offsetof(struct libeth_xdp_buff, desc) ==
|
|
offsetof(struct xdp_buff_xsk, cb));
|
|
static_assert(IS_ALIGNED(sizeof(struct xdp_buff_xsk),
|
|
__alignof(struct libeth_xdp_buff)));
|
|
|
|
/**
|
|
* __LIBETH_XDP_ONSTACK_BUFF - declare a &libeth_xdp_buff on the stack
|
|
* @name: name of the variable to declare
|
|
* @...: sizeof() of the driver-private data
|
|
*/
|
|
#define __LIBETH_XDP_ONSTACK_BUFF(name, ...) \
|
|
___LIBETH_XDP_ONSTACK_BUFF(name, ##__VA_ARGS__)
|
|
/**
|
|
* LIBETH_XDP_ONSTACK_BUFF - declare a &libeth_xdp_buff on the stack
|
|
* @name: name of the variable to declare
|
|
* @...: type or variable name of the driver-private data
|
|
*/
|
|
#define LIBETH_XDP_ONSTACK_BUFF(name, ...) \
|
|
__LIBETH_XDP_ONSTACK_BUFF(name, __libeth_xdp_priv_sz(__VA_ARGS__))
|
|
|
|
#define ___LIBETH_XDP_ONSTACK_BUFF(name, ...) \
|
|
__DEFINE_FLEX(struct libeth_xdp_buff, name, priv, \
|
|
LIBETH_XDP_PRIV_SZ(__VA_ARGS__ + 0), \
|
|
__uninitialized); \
|
|
LIBETH_XDP_ASSERT_PRIV_SZ(__VA_ARGS__ + 0)
|
|
|
|
#define __libeth_xdp_priv_sz(...) \
|
|
CONCATENATE(__libeth_xdp_psz, COUNT_ARGS(__VA_ARGS__))(__VA_ARGS__)
|
|
|
|
#define __libeth_xdp_psz0(...)
|
|
#define __libeth_xdp_psz1(...) sizeof(__VA_ARGS__)
|
|
|
|
#define LIBETH_XDP_PRIV_SZ(sz) \
|
|
(ALIGN(sz, __alignof(struct libeth_xdp_buff)) / sizeof(long))
|
|
|
|
/* Performs XSK_CHECK_PRIV_TYPE() */
|
|
#define LIBETH_XDP_ASSERT_PRIV_SZ(sz) \
|
|
static_assert(offsetofend(struct xdp_buff_xsk, cb) >= \
|
|
struct_size_t(struct libeth_xdp_buff, priv, \
|
|
LIBETH_XDP_PRIV_SZ(sz)))
|
|
|
|
/* XDPSQ sharing */
|
|
|
|
DECLARE_STATIC_KEY_FALSE(libeth_xdpsq_share);
|
|
|
|
/**
|
|
* libeth_xdpsq_num - calculate optimal number of XDPSQs for this device + sys
|
|
* @rxq: current number of active Rx queues
|
|
* @txq: current number of active Tx queues
|
|
* @max: maximum number of Tx queues
|
|
*
|
|
* Each RQ must have its own XDPSQ for XSk pairs, each CPU must have own XDPSQ
|
|
* for lockless sending (``XDP_TX``, .ndo_xdp_xmit()). Cap the maximum of these
|
|
* two with the number of SQs the device can have (minus used ones).
|
|
*
|
|
* Return: number of XDP Tx queues the device needs to use.
|
|
*/
|
|
static inline u32 libeth_xdpsq_num(u32 rxq, u32 txq, u32 max)
|
|
{
|
|
return min(max(nr_cpu_ids, rxq), max - txq);
|
|
}
|
|
|
|
/**
|
|
* libeth_xdpsq_shared - whether XDPSQs can be shared between several CPUs
|
|
* @num: number of active XDPSQs
|
|
*
|
|
* Return: true if there's no 1:1 XDPSQ/CPU association, false otherwise.
|
|
*/
|
|
static inline bool libeth_xdpsq_shared(u32 num)
|
|
{
|
|
return num < nr_cpu_ids;
|
|
}
|
|
|
|
/**
|
|
* libeth_xdpsq_id - get XDPSQ index corresponding to this CPU
|
|
* @num: number of active XDPSQs
|
|
*
|
|
* Helper for libeth_xdp routines, do not use in drivers directly.
|
|
*
|
|
* Return: XDPSQ index needs to be used on this CPU.
|
|
*/
|
|
static inline u32 libeth_xdpsq_id(u32 num)
|
|
{
|
|
u32 ret = raw_smp_processor_id();
|
|
|
|
if (static_branch_unlikely(&libeth_xdpsq_share) &&
|
|
libeth_xdpsq_shared(num))
|
|
ret %= num;
|
|
|
|
return ret;
|
|
}
|
|
|
|
void __libeth_xdpsq_get(struct libeth_xdpsq_lock *lock,
|
|
const struct net_device *dev);
|
|
void __libeth_xdpsq_put(struct libeth_xdpsq_lock *lock,
|
|
const struct net_device *dev);
|
|
|
|
/**
|
|
* libeth_xdpsq_get - initialize &libeth_xdpsq_lock
|
|
* @lock: lock to initialize
|
|
* @dev: netdev which this lock belongs to
|
|
* @share: whether XDPSQs can be shared
|
|
*
|
|
* Tracks the current XDPSQ association and enables the static lock
|
|
* if needed.
|
|
*/
|
|
static inline void libeth_xdpsq_get(struct libeth_xdpsq_lock *lock,
|
|
const struct net_device *dev,
|
|
bool share)
|
|
{
|
|
if (unlikely(share))
|
|
__libeth_xdpsq_get(lock, dev);
|
|
}
|
|
|
|
/**
|
|
* libeth_xdpsq_put - deinitialize &libeth_xdpsq_lock
|
|
* @lock: lock to deinitialize
|
|
* @dev: netdev which this lock belongs to
|
|
*
|
|
* Tracks the current XDPSQ association and disables the static lock
|
|
* if needed.
|
|
*/
|
|
static inline void libeth_xdpsq_put(struct libeth_xdpsq_lock *lock,
|
|
const struct net_device *dev)
|
|
{
|
|
if (static_branch_unlikely(&libeth_xdpsq_share) && lock->share)
|
|
__libeth_xdpsq_put(lock, dev);
|
|
}
|
|
|
|
void __libeth_xdpsq_lock(struct libeth_xdpsq_lock *lock);
|
|
void __libeth_xdpsq_unlock(struct libeth_xdpsq_lock *lock);
|
|
|
|
/**
|
|
* libeth_xdpsq_lock - grab &libeth_xdpsq_lock if needed
|
|
* @lock: lock to take
|
|
*
|
|
* Touches the underlying spinlock only if the static key is enabled
|
|
* and the queue itself is marked as shareable.
|
|
*/
|
|
static inline void libeth_xdpsq_lock(struct libeth_xdpsq_lock *lock)
|
|
{
|
|
if (static_branch_unlikely(&libeth_xdpsq_share) && lock->share)
|
|
__libeth_xdpsq_lock(lock);
|
|
}
|
|
|
|
/**
|
|
* libeth_xdpsq_unlock - free &libeth_xdpsq_lock if needed
|
|
* @lock: lock to free
|
|
*
|
|
* Touches the underlying spinlock only if the static key is enabled
|
|
* and the queue itself is marked as shareable.
|
|
*/
|
|
static inline void libeth_xdpsq_unlock(struct libeth_xdpsq_lock *lock)
|
|
{
|
|
if (static_branch_unlikely(&libeth_xdpsq_share) && lock->share)
|
|
__libeth_xdpsq_unlock(lock);
|
|
}
|
|
|
|
/* XDPSQ clean-up timers */
|
|
|
|
void libeth_xdpsq_init_timer(struct libeth_xdpsq_timer *timer, void *xdpsq,
|
|
struct libeth_xdpsq_lock *lock,
|
|
void (*poll)(struct work_struct *work));
|
|
|
|
/**
|
|
* libeth_xdpsq_deinit_timer - deinitialize &libeth_xdpsq_timer
|
|
* @timer: timer to deinitialize
|
|
*
|
|
* Flush and disable the underlying workqueue.
|
|
*/
|
|
static inline void libeth_xdpsq_deinit_timer(struct libeth_xdpsq_timer *timer)
|
|
{
|
|
cancel_delayed_work_sync(&timer->dwork);
|
|
}
|
|
|
|
/**
|
|
* libeth_xdpsq_queue_timer - run &libeth_xdpsq_timer
|
|
* @timer: timer to queue
|
|
*
|
|
* Should be called after the queue was filled and the transmission was run
|
|
* to complete the pending buffers if no further sending will be done in a
|
|
* second (-> lazy cleaning won't happen).
|
|
* If the timer was already run, it will be requeued back to one second
|
|
* timeout again.
|
|
*/
|
|
static inline void libeth_xdpsq_queue_timer(struct libeth_xdpsq_timer *timer)
|
|
{
|
|
mod_delayed_work_on(raw_smp_processor_id(), system_bh_highpri_wq,
|
|
&timer->dwork, HZ);
|
|
}
|
|
|
|
/**
|
|
* libeth_xdpsq_run_timer - wrapper to run a queue clean-up on a timer event
|
|
* @work: workqueue belonging to the corresponding timer
|
|
* @poll: driver-specific completion queue poll function
|
|
*
|
|
* Run the polling function on the locked queue and requeue the timer if
|
|
* there's more work to do.
|
|
* Designed to be used via LIBETH_XDP_DEFINE_TIMER() below.
|
|
*/
|
|
static __always_inline void
|
|
libeth_xdpsq_run_timer(struct work_struct *work,
|
|
u32 (*poll)(void *xdpsq, u32 budget))
|
|
{
|
|
struct libeth_xdpsq_timer *timer = container_of(work, typeof(*timer),
|
|
dwork.work);
|
|
|
|
libeth_xdpsq_lock(timer->lock);
|
|
|
|
if (poll(timer->xdpsq, U32_MAX))
|
|
libeth_xdpsq_queue_timer(timer);
|
|
|
|
libeth_xdpsq_unlock(timer->lock);
|
|
}
|
|
|
|
/* Common Tx bits */
|
|
|
|
/**
|
|
* enum - libeth_xdp internal Tx flags
|
|
* @LIBETH_XDP_TX_BULK: one bulk size at which it will be flushed to the queue
|
|
* @LIBETH_XDP_TX_BATCH: batch size for which the queue fill loop is unrolled
|
|
* @LIBETH_XDP_TX_DROP: indicates the send function must drop frames not sent
|
|
* @LIBETH_XDP_TX_NDO: whether the send function is called from .ndo_xdp_xmit()
|
|
* @LIBETH_XDP_TX_XSK: whether the function is called for ``XDP_TX`` for XSk
|
|
*/
|
|
enum {
|
|
LIBETH_XDP_TX_BULK = DEV_MAP_BULK_SIZE,
|
|
LIBETH_XDP_TX_BATCH = 8,
|
|
|
|
LIBETH_XDP_TX_DROP = BIT(0),
|
|
LIBETH_XDP_TX_NDO = BIT(1),
|
|
LIBETH_XDP_TX_XSK = BIT(2),
|
|
};
|
|
|
|
/**
|
|
* enum - &libeth_xdp_tx_frame and &libeth_xdp_tx_desc flags
|
|
* @LIBETH_XDP_TX_LEN: only for ``XDP_TX``, [15:0] of ::len_fl is actual length
|
|
* @LIBETH_XDP_TX_CSUM: for XSk xmit, enable checksum offload
|
|
* @LIBETH_XDP_TX_XSKMD: for XSk xmit, mask of the metadata bits
|
|
* @LIBETH_XDP_TX_FIRST: indicates the frag is the first one of the frame
|
|
* @LIBETH_XDP_TX_LAST: whether the frag is the last one of the frame
|
|
* @LIBETH_XDP_TX_MULTI: whether the frame contains several frags
|
|
* @LIBETH_XDP_TX_FLAGS: only for ``XDP_TX``, [31:16] of ::len_fl is flags
|
|
*/
|
|
enum {
|
|
LIBETH_XDP_TX_LEN = GENMASK(15, 0),
|
|
|
|
LIBETH_XDP_TX_CSUM = XDP_TXMD_FLAGS_CHECKSUM,
|
|
LIBETH_XDP_TX_XSKMD = LIBETH_XDP_TX_LEN,
|
|
|
|
LIBETH_XDP_TX_FIRST = BIT(16),
|
|
LIBETH_XDP_TX_LAST = BIT(17),
|
|
LIBETH_XDP_TX_MULTI = BIT(18),
|
|
|
|
LIBETH_XDP_TX_FLAGS = GENMASK(31, 16),
|
|
};
|
|
|
|
/**
|
|
* struct libeth_xdp_tx_frame - represents one XDP Tx element
|
|
* @data: frame start pointer for ``XDP_TX``
|
|
* @len_fl: ``XDP_TX``, combined flags [31:16] and len [15:0] field for speed
|
|
* @soff: ``XDP_TX``, offset from @data to the start of &skb_shared_info
|
|
* @frag: one (non-head) frag for ``XDP_TX``
|
|
* @xdpf: &xdp_frame for the head frag for .ndo_xdp_xmit()
|
|
* @dma: DMA address of the non-head frag for .ndo_xdp_xmit()
|
|
* @xsk: ``XDP_TX`` for XSk, XDP buffer for any frag
|
|
* @len: frag length for XSk ``XDP_TX`` and .ndo_xdp_xmit()
|
|
* @flags: Tx flags for the above
|
|
* @opts: combined @len + @flags for the above for speed
|
|
* @desc: XSk xmit descriptor for direct casting
|
|
*/
|
|
struct libeth_xdp_tx_frame {
|
|
union {
|
|
/* ``XDP_TX`` */
|
|
struct {
|
|
void *data;
|
|
u32 len_fl;
|
|
u32 soff;
|
|
};
|
|
|
|
/* ``XDP_TX`` frag */
|
|
skb_frag_t frag;
|
|
|
|
/* .ndo_xdp_xmit(), XSk ``XDP_TX`` */
|
|
struct {
|
|
union {
|
|
struct xdp_frame *xdpf;
|
|
dma_addr_t dma;
|
|
|
|
struct libeth_xdp_buff *xsk;
|
|
};
|
|
union {
|
|
struct {
|
|
u32 len;
|
|
u32 flags;
|
|
};
|
|
aligned_u64 opts;
|
|
};
|
|
};
|
|
|
|
/* XSk xmit */
|
|
struct xdp_desc desc;
|
|
};
|
|
} __aligned(sizeof(struct xdp_desc));
|
|
static_assert(offsetof(struct libeth_xdp_tx_frame, frag.len) ==
|
|
offsetof(struct libeth_xdp_tx_frame, len_fl));
|
|
static_assert(sizeof(struct libeth_xdp_tx_frame) == sizeof(struct xdp_desc));
|
|
|
|
/**
|
|
* struct libeth_xdp_tx_bulk - XDP Tx frame bulk for bulk sending
|
|
* @prog: corresponding active XDP program, %NULL for .ndo_xdp_xmit()
|
|
* @dev: &net_device which the frames are transmitted on
|
|
* @xdpsq: shortcut to the corresponding driver-specific XDPSQ structure
|
|
* @act_mask: Rx only, mask of all the XDP prog verdicts for that NAPI session
|
|
* @count: current number of frames in @bulk
|
|
* @bulk: array of queued frames for bulk Tx
|
|
*
|
|
* All XDP Tx operations except XSk xmit queue each frame to the bulk first
|
|
* and flush it when @count reaches the array end. Bulk is always placed on
|
|
* the stack for performance. One bulk element contains all the data necessary
|
|
* for sending a frame and then freeing it on completion.
|
|
* For XSk xmit, Tx descriptor array from &xsk_buff_pool is casted directly
|
|
* to &libeth_xdp_tx_frame as they are compatible and the bulk structure is
|
|
* not used.
|
|
*/
|
|
struct libeth_xdp_tx_bulk {
|
|
const struct bpf_prog *prog;
|
|
struct net_device *dev;
|
|
void *xdpsq;
|
|
|
|
u32 act_mask;
|
|
u32 count;
|
|
struct libeth_xdp_tx_frame bulk[LIBETH_XDP_TX_BULK];
|
|
} __aligned(sizeof(struct libeth_xdp_tx_frame));
|
|
|
|
/**
|
|
* LIBETH_XDP_ONSTACK_BULK - declare &libeth_xdp_tx_bulk on the stack
|
|
* @bq: name of the variable to declare
|
|
*
|
|
* Helper to declare a bulk on the stack with a compiler hint that it should
|
|
* not be initialized automatically (with `CONFIG_INIT_STACK_ALL_*`) for
|
|
* performance reasons.
|
|
*/
|
|
#define LIBETH_XDP_ONSTACK_BULK(bq) \
|
|
struct libeth_xdp_tx_bulk bq __uninitialized
|
|
|
|
/**
|
|
* struct libeth_xdpsq - abstraction for an XDPSQ
|
|
* @pool: XSk buffer pool for XSk ``XDP_TX`` and xmit
|
|
* @sqes: array of Tx buffers from the actual queue struct
|
|
* @descs: opaque pointer to the HW descriptor array
|
|
* @ntu: pointer to the next free descriptor index
|
|
* @count: number of descriptors on that queue
|
|
* @pending: pointer to the number of sent-not-completed descs on that queue
|
|
* @xdp_tx: pointer to the above, but only for non-XSk-xmit frames
|
|
* @lock: corresponding XDPSQ lock
|
|
*
|
|
* Abstraction for driver-independent implementation of Tx. Placed on the stack
|
|
* and filled by the driver before the transmission, so that the generic
|
|
* functions can access and modify driver-specific resources.
|
|
*/
|
|
struct libeth_xdpsq {
|
|
struct xsk_buff_pool *pool;
|
|
struct libeth_sqe *sqes;
|
|
void *descs;
|
|
|
|
u32 *ntu;
|
|
u32 count;
|
|
|
|
u32 *pending;
|
|
u32 *xdp_tx;
|
|
struct libeth_xdpsq_lock *lock;
|
|
};
|
|
|
|
/**
|
|
* struct libeth_xdp_tx_desc - abstraction for an XDP Tx descriptor
|
|
* @addr: DMA address of the frame
|
|
* @len: length of the frame
|
|
* @flags: XDP Tx flags
|
|
* @opts: combined @len + @flags for speed
|
|
*
|
|
* Filled by the generic functions and then passed to driver-specific functions
|
|
* to fill a HW Tx descriptor, always placed on the [function] stack.
|
|
*/
|
|
struct libeth_xdp_tx_desc {
|
|
dma_addr_t addr;
|
|
union {
|
|
struct {
|
|
u32 len;
|
|
u32 flags;
|
|
};
|
|
aligned_u64 opts;
|
|
};
|
|
} __aligned_largest;
|
|
|
|
/**
|
|
* libeth_xdp_ptr_to_priv - convert pointer to a libeth_xdp u64 priv
|
|
* @ptr: pointer to convert
|
|
*
|
|
* The main sending function passes private data as the largest scalar, u64.
|
|
* Use this helper when you want to pass a pointer there.
|
|
*/
|
|
#define libeth_xdp_ptr_to_priv(ptr) ({ \
|
|
typecheck_pointer(ptr); \
|
|
((u64)(uintptr_t)(ptr)); \
|
|
})
|
|
/**
|
|
* libeth_xdp_priv_to_ptr - convert libeth_xdp u64 priv to a pointer
|
|
* @priv: private data to convert
|
|
*
|
|
* The main sending function passes private data as the largest scalar, u64.
|
|
* Use this helper when your callback takes this u64 and you want to convert
|
|
* it back to a pointer.
|
|
*/
|
|
#define libeth_xdp_priv_to_ptr(priv) ({ \
|
|
static_assert(__same_type(priv, u64)); \
|
|
((const void *)(uintptr_t)(priv)); \
|
|
})
|
|
|
|
/*
|
|
* On 64-bit systems, assigning one u64 is faster than two u32s. When ::len
|
|
* occupies lowest 32 bits (LE), whole ::opts can be assigned directly instead.
|
|
*/
|
|
#ifdef __LITTLE_ENDIAN
|
|
#define __LIBETH_WORD_ACCESS 1
|
|
#endif
|
|
#ifdef __LIBETH_WORD_ACCESS
|
|
#define __libeth_xdp_tx_len(flen, ...) \
|
|
.opts = ((flen) | FIELD_PREP(GENMASK_ULL(63, 32), (__VA_ARGS__ + 0)))
|
|
#else
|
|
#define __libeth_xdp_tx_len(flen, ...) \
|
|
.len = (flen), .flags = (__VA_ARGS__ + 0)
|
|
#endif
|
|
|
|
/**
|
|
* libeth_xdp_tx_xmit_bulk - main XDP Tx function
|
|
* @bulk: array of frames to send
|
|
* @xdpsq: pointer to the driver-specific XDPSQ struct
|
|
* @n: number of frames to send
|
|
* @unroll: whether to unroll the queue filling loop for speed
|
|
* @priv: driver-specific private data
|
|
* @prep: callback for cleaning the queue and filling abstract &libeth_xdpsq
|
|
* @fill: internal callback for filling &libeth_sqe and &libeth_xdp_tx_desc
|
|
* @xmit: callback for filling a HW descriptor with the frame info
|
|
*
|
|
* Internal abstraction for placing @n XDP Tx frames on the HW XDPSQ. Used for
|
|
* all types of frames: ``XDP_TX``, .ndo_xdp_xmit(), XSk ``XDP_TX``, and XSk
|
|
* xmit.
|
|
* @prep must lock the queue as this function releases it at the end. @unroll
|
|
* greatly increases the object code size, but also greatly increases XSk xmit
|
|
* performance; for other types of frames, it's not enabled.
|
|
* The compilers inline all those onstack abstractions to direct data accesses.
|
|
*
|
|
* Return: number of frames actually placed on the queue, <= @n. The function
|
|
* can't fail, but can send less frames if there's no enough free descriptors
|
|
* available. The actual free space is returned by @prep from the driver.
|
|
*/
|
|
static __always_inline u32
|
|
libeth_xdp_tx_xmit_bulk(const struct libeth_xdp_tx_frame *bulk, void *xdpsq,
|
|
u32 n, bool unroll, u64 priv,
|
|
u32 (*prep)(void *xdpsq, struct libeth_xdpsq *sq),
|
|
struct libeth_xdp_tx_desc
|
|
(*fill)(struct libeth_xdp_tx_frame frm, u32 i,
|
|
const struct libeth_xdpsq *sq, u64 priv),
|
|
void (*xmit)(struct libeth_xdp_tx_desc desc, u32 i,
|
|
const struct libeth_xdpsq *sq, u64 priv))
|
|
{
|
|
struct libeth_xdpsq sq __uninitialized;
|
|
u32 this, batched, off = 0;
|
|
u32 ntu, i = 0;
|
|
|
|
n = min(n, prep(xdpsq, &sq));
|
|
if (unlikely(!n))
|
|
goto unlock;
|
|
|
|
ntu = *sq.ntu;
|
|
|
|
this = sq.count - ntu;
|
|
if (likely(this > n))
|
|
this = n;
|
|
|
|
again:
|
|
if (!unroll)
|
|
goto linear;
|
|
|
|
batched = ALIGN_DOWN(this, LIBETH_XDP_TX_BATCH);
|
|
|
|
for ( ; i < off + batched; i += LIBETH_XDP_TX_BATCH) {
|
|
u32 base = ntu + i - off;
|
|
|
|
unrolled_count(LIBETH_XDP_TX_BATCH)
|
|
for (u32 j = 0; j < LIBETH_XDP_TX_BATCH; j++)
|
|
xmit(fill(bulk[i + j], base + j, &sq, priv),
|
|
base + j, &sq, priv);
|
|
}
|
|
|
|
if (batched < this) {
|
|
linear:
|
|
for ( ; i < off + this; i++)
|
|
xmit(fill(bulk[i], ntu + i - off, &sq, priv),
|
|
ntu + i - off, &sq, priv);
|
|
}
|
|
|
|
ntu += this;
|
|
if (likely(ntu < sq.count))
|
|
goto out;
|
|
|
|
ntu = 0;
|
|
|
|
if (i < n) {
|
|
this = n - i;
|
|
off = i;
|
|
|
|
goto again;
|
|
}
|
|
|
|
out:
|
|
*sq.ntu = ntu;
|
|
*sq.pending += n;
|
|
if (sq.xdp_tx)
|
|
*sq.xdp_tx += n;
|
|
|
|
unlock:
|
|
libeth_xdpsq_unlock(sq.lock);
|
|
|
|
return n;
|
|
}
|
|
|
|
/* ``XDP_TX`` bulking */
|
|
|
|
void libeth_xdp_return_buff_slow(struct libeth_xdp_buff *xdp);
|
|
|
|
/**
|
|
* libeth_xdp_tx_queue_head - internal helper for queueing one ``XDP_TX`` head
|
|
* @bq: XDP Tx bulk to queue the head frag to
|
|
* @xdp: XDP buffer with the head to queue
|
|
*
|
|
* Return: false if it's the only frag of the frame, true if it's an S/G frame.
|
|
*/
|
|
static inline bool libeth_xdp_tx_queue_head(struct libeth_xdp_tx_bulk *bq,
|
|
const struct libeth_xdp_buff *xdp)
|
|
{
|
|
const struct xdp_buff *base = &xdp->base;
|
|
|
|
bq->bulk[bq->count++] = (typeof(*bq->bulk)){
|
|
.data = xdp->data,
|
|
.len_fl = (base->data_end - xdp->data) | LIBETH_XDP_TX_FIRST,
|
|
.soff = xdp_data_hard_end(base) - xdp->data,
|
|
};
|
|
|
|
if (!xdp_buff_has_frags(base))
|
|
return false;
|
|
|
|
bq->bulk[bq->count - 1].len_fl |= LIBETH_XDP_TX_MULTI;
|
|
|
|
return true;
|
|
}
|
|
|
|
/**
|
|
* libeth_xdp_tx_queue_frag - internal helper for queueing one ``XDP_TX`` frag
|
|
* @bq: XDP Tx bulk to queue the frag to
|
|
* @frag: frag to queue
|
|
*/
|
|
static inline void libeth_xdp_tx_queue_frag(struct libeth_xdp_tx_bulk *bq,
|
|
const skb_frag_t *frag)
|
|
{
|
|
bq->bulk[bq->count++].frag = *frag;
|
|
}
|
|
|
|
/**
|
|
* libeth_xdp_tx_queue_bulk - internal helper for queueing one ``XDP_TX`` frame
|
|
* @bq: XDP Tx bulk to queue the frame to
|
|
* @xdp: XDP buffer to queue
|
|
* @flush_bulk: driver callback to flush the bulk to the HW queue
|
|
*
|
|
* Return: true on success, false on flush error.
|
|
*/
|
|
static __always_inline bool
|
|
libeth_xdp_tx_queue_bulk(struct libeth_xdp_tx_bulk *bq,
|
|
struct libeth_xdp_buff *xdp,
|
|
bool (*flush_bulk)(struct libeth_xdp_tx_bulk *bq,
|
|
u32 flags))
|
|
{
|
|
const struct skb_shared_info *sinfo;
|
|
bool ret = true;
|
|
u32 nr_frags;
|
|
|
|
if (unlikely(bq->count == LIBETH_XDP_TX_BULK) &&
|
|
unlikely(!flush_bulk(bq, 0))) {
|
|
libeth_xdp_return_buff_slow(xdp);
|
|
return false;
|
|
}
|
|
|
|
if (!libeth_xdp_tx_queue_head(bq, xdp))
|
|
goto out;
|
|
|
|
sinfo = xdp_get_shared_info_from_buff(&xdp->base);
|
|
nr_frags = sinfo->nr_frags;
|
|
|
|
for (u32 i = 0; i < nr_frags; i++) {
|
|
if (unlikely(bq->count == LIBETH_XDP_TX_BULK) &&
|
|
unlikely(!flush_bulk(bq, 0))) {
|
|
ret = false;
|
|
break;
|
|
}
|
|
|
|
libeth_xdp_tx_queue_frag(bq, &sinfo->frags[i]);
|
|
}
|
|
|
|
out:
|
|
bq->bulk[bq->count - 1].len_fl |= LIBETH_XDP_TX_LAST;
|
|
xdp->data = NULL;
|
|
|
|
return ret;
|
|
}
|
|
|
|
/**
|
|
* libeth_xdp_tx_fill_stats - fill &libeth_sqe with ``XDP_TX`` frame stats
|
|
* @sqe: SQ element to fill
|
|
* @desc: libeth_xdp Tx descriptor
|
|
* @sinfo: &skb_shared_info for this frame
|
|
*
|
|
* Internal helper for filling an SQE with the frame stats, do not use in
|
|
* drivers. Fills the number of frags and bytes for this frame.
|
|
*/
|
|
#define libeth_xdp_tx_fill_stats(sqe, desc, sinfo) \
|
|
__libeth_xdp_tx_fill_stats(sqe, desc, sinfo, __UNIQUE_ID(sqe_), \
|
|
__UNIQUE_ID(desc_), __UNIQUE_ID(sinfo_))
|
|
|
|
#define __libeth_xdp_tx_fill_stats(sqe, desc, sinfo, ue, ud, us) do { \
|
|
const struct libeth_xdp_tx_desc *ud = (desc); \
|
|
const struct skb_shared_info *us; \
|
|
struct libeth_sqe *ue = (sqe); \
|
|
\
|
|
ue->nr_frags = 1; \
|
|
ue->bytes = ud->len; \
|
|
\
|
|
if (ud->flags & LIBETH_XDP_TX_MULTI) { \
|
|
us = (sinfo); \
|
|
ue->nr_frags += us->nr_frags; \
|
|
ue->bytes += us->xdp_frags_size; \
|
|
} \
|
|
} while (0)
|
|
|
|
/**
|
|
* libeth_xdp_tx_fill_buf - internal helper to fill one ``XDP_TX`` &libeth_sqe
|
|
* @frm: XDP Tx frame from the bulk
|
|
* @i: index on the HW queue
|
|
* @sq: XDPSQ abstraction for the queue
|
|
* @priv: private data
|
|
*
|
|
* Return: XDP Tx descriptor with the synced DMA and other info to pass to
|
|
* the driver callback.
|
|
*/
|
|
static inline struct libeth_xdp_tx_desc
|
|
libeth_xdp_tx_fill_buf(struct libeth_xdp_tx_frame frm, u32 i,
|
|
const struct libeth_xdpsq *sq, u64 priv)
|
|
{
|
|
struct libeth_xdp_tx_desc desc;
|
|
struct skb_shared_info *sinfo;
|
|
skb_frag_t *frag = &frm.frag;
|
|
struct libeth_sqe *sqe;
|
|
netmem_ref netmem;
|
|
|
|
if (frm.len_fl & LIBETH_XDP_TX_FIRST) {
|
|
sinfo = frm.data + frm.soff;
|
|
skb_frag_fill_netmem_desc(frag, virt_to_netmem(frm.data),
|
|
offset_in_page(frm.data),
|
|
frm.len_fl);
|
|
} else {
|
|
sinfo = NULL;
|
|
}
|
|
|
|
netmem = skb_frag_netmem(frag);
|
|
desc = (typeof(desc)){
|
|
.addr = page_pool_get_dma_addr_netmem(netmem) +
|
|
skb_frag_off(frag),
|
|
.len = skb_frag_size(frag) & LIBETH_XDP_TX_LEN,
|
|
.flags = skb_frag_size(frag) & LIBETH_XDP_TX_FLAGS,
|
|
};
|
|
|
|
dma_sync_single_for_device(__netmem_get_pp(netmem)->p.dev, desc.addr,
|
|
desc.len, DMA_BIDIRECTIONAL);
|
|
|
|
if (!sinfo)
|
|
return desc;
|
|
|
|
sqe = &sq->sqes[i];
|
|
sqe->type = LIBETH_SQE_XDP_TX;
|
|
sqe->sinfo = sinfo;
|
|
libeth_xdp_tx_fill_stats(sqe, &desc, sinfo);
|
|
|
|
return desc;
|
|
}
|
|
|
|
void libeth_xdp_tx_exception(struct libeth_xdp_tx_bulk *bq, u32 sent,
|
|
u32 flags);
|
|
|
|
/**
|
|
* __libeth_xdp_tx_flush_bulk - internal helper to flush one XDP Tx bulk
|
|
* @bq: bulk to flush
|
|
* @flags: XDP TX flags (.ndo_xdp_xmit(), XSk etc.)
|
|
* @prep: driver-specific callback to prepare the queue for sending
|
|
* @fill: libeth_xdp callback to fill &libeth_sqe and &libeth_xdp_tx_desc
|
|
* @xmit: driver callback to fill a HW descriptor
|
|
*
|
|
* Internal abstraction to create bulk flush functions for drivers. Used for
|
|
* everything except XSk xmit.
|
|
*
|
|
* Return: true if anything was sent, false otherwise.
|
|
*/
|
|
static __always_inline bool
|
|
__libeth_xdp_tx_flush_bulk(struct libeth_xdp_tx_bulk *bq, u32 flags,
|
|
u32 (*prep)(void *xdpsq, struct libeth_xdpsq *sq),
|
|
struct libeth_xdp_tx_desc
|
|
(*fill)(struct libeth_xdp_tx_frame frm, u32 i,
|
|
const struct libeth_xdpsq *sq, u64 priv),
|
|
void (*xmit)(struct libeth_xdp_tx_desc desc, u32 i,
|
|
const struct libeth_xdpsq *sq,
|
|
u64 priv))
|
|
{
|
|
u32 sent, drops;
|
|
int err = 0;
|
|
|
|
sent = libeth_xdp_tx_xmit_bulk(bq->bulk, bq->xdpsq,
|
|
min(bq->count, LIBETH_XDP_TX_BULK),
|
|
false, 0, prep, fill, xmit);
|
|
drops = bq->count - sent;
|
|
|
|
if (unlikely(drops)) {
|
|
libeth_xdp_tx_exception(bq, sent, flags);
|
|
err = -ENXIO;
|
|
} else {
|
|
bq->count = 0;
|
|
}
|
|
|
|
trace_xdp_bulk_tx(bq->dev, sent, drops, err);
|
|
|
|
return likely(sent);
|
|
}
|
|
|
|
/**
|
|
* libeth_xdp_tx_flush_bulk - wrapper to define flush of one ``XDP_TX`` bulk
|
|
* @bq: bulk to flush
|
|
* @flags: Tx flags, see above
|
|
* @prep: driver callback to prepare the queue
|
|
* @xmit: driver callback to fill a HW descriptor
|
|
*
|
|
* Use via LIBETH_XDP_DEFINE_FLUSH_TX() to define an ``XDP_TX`` driver
|
|
* callback.
|
|
*/
|
|
#define libeth_xdp_tx_flush_bulk(bq, flags, prep, xmit) \
|
|
__libeth_xdp_tx_flush_bulk(bq, flags, prep, libeth_xdp_tx_fill_buf, \
|
|
xmit)
|
|
|
|
/* .ndo_xdp_xmit() implementation */
|
|
|
|
/**
|
|
* libeth_xdp_xmit_init_bulk - internal helper to initialize bulk for XDP xmit
|
|
* @bq: bulk to initialize
|
|
* @dev: target &net_device
|
|
* @xdpsqs: array of driver-specific XDPSQ structs
|
|
* @num: number of active XDPSQs (the above array length)
|
|
*/
|
|
#define libeth_xdp_xmit_init_bulk(bq, dev, xdpsqs, num) \
|
|
__libeth_xdp_xmit_init_bulk(bq, dev, (xdpsqs)[libeth_xdpsq_id(num)])
|
|
|
|
static inline void __libeth_xdp_xmit_init_bulk(struct libeth_xdp_tx_bulk *bq,
|
|
struct net_device *dev,
|
|
void *xdpsq)
|
|
{
|
|
bq->dev = dev;
|
|
bq->xdpsq = xdpsq;
|
|
bq->count = 0;
|
|
}
|
|
|
|
/**
|
|
* libeth_xdp_xmit_frame_dma - internal helper to access DMA of an &xdp_frame
|
|
* @xf: pointer to the XDP frame
|
|
*
|
|
* There's no place in &libeth_xdp_tx_frame to store DMA address for an
|
|
* &xdp_frame head. The headroom is used then, the address is placed right
|
|
* after the frame struct, naturally aligned.
|
|
*
|
|
* Return: pointer to the DMA address to use.
|
|
*/
|
|
#define libeth_xdp_xmit_frame_dma(xf) \
|
|
_Generic((xf), \
|
|
const struct xdp_frame *: \
|
|
(const dma_addr_t *)__libeth_xdp_xmit_frame_dma(xf), \
|
|
struct xdp_frame *: \
|
|
(dma_addr_t *)__libeth_xdp_xmit_frame_dma(xf) \
|
|
)
|
|
|
|
static inline void *__libeth_xdp_xmit_frame_dma(const struct xdp_frame *xdpf)
|
|
{
|
|
void *addr = (void *)(xdpf + 1);
|
|
|
|
if (!IS_ENABLED(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) &&
|
|
__alignof(*xdpf) < sizeof(dma_addr_t))
|
|
addr = PTR_ALIGN(addr, sizeof(dma_addr_t));
|
|
|
|
return addr;
|
|
}
|
|
|
|
/**
|
|
* libeth_xdp_xmit_queue_head - internal helper for queueing one XDP xmit head
|
|
* @bq: XDP Tx bulk to queue the head frag to
|
|
* @xdpf: XDP frame with the head to queue
|
|
* @dev: device to perform DMA mapping
|
|
*
|
|
* Return: ``LIBETH_XDP_DROP`` on DMA mapping error,
|
|
* ``LIBETH_XDP_PASS`` if it's the only frag in the frame,
|
|
* ``LIBETH_XDP_TX`` if it's an S/G frame.
|
|
*/
|
|
static inline u32 libeth_xdp_xmit_queue_head(struct libeth_xdp_tx_bulk *bq,
|
|
struct xdp_frame *xdpf,
|
|
struct device *dev)
|
|
{
|
|
dma_addr_t dma;
|
|
|
|
dma = dma_map_single(dev, xdpf->data, xdpf->len, DMA_TO_DEVICE);
|
|
if (dma_mapping_error(dev, dma))
|
|
return LIBETH_XDP_DROP;
|
|
|
|
*libeth_xdp_xmit_frame_dma(xdpf) = dma;
|
|
|
|
bq->bulk[bq->count++] = (typeof(*bq->bulk)){
|
|
.xdpf = xdpf,
|
|
__libeth_xdp_tx_len(xdpf->len, LIBETH_XDP_TX_FIRST),
|
|
};
|
|
|
|
if (!xdp_frame_has_frags(xdpf))
|
|
return LIBETH_XDP_PASS;
|
|
|
|
bq->bulk[bq->count - 1].flags |= LIBETH_XDP_TX_MULTI;
|
|
|
|
return LIBETH_XDP_TX;
|
|
}
|
|
|
|
/**
|
|
* libeth_xdp_xmit_queue_frag - internal helper for queueing one XDP xmit frag
|
|
* @bq: XDP Tx bulk to queue the frag to
|
|
* @frag: frag to queue
|
|
* @dev: device to perform DMA mapping
|
|
*
|
|
* Return: true on success, false on DMA mapping error.
|
|
*/
|
|
static inline bool libeth_xdp_xmit_queue_frag(struct libeth_xdp_tx_bulk *bq,
|
|
const skb_frag_t *frag,
|
|
struct device *dev)
|
|
{
|
|
dma_addr_t dma;
|
|
|
|
dma = skb_frag_dma_map(dev, frag);
|
|
if (dma_mapping_error(dev, dma))
|
|
return false;
|
|
|
|
bq->bulk[bq->count++] = (typeof(*bq->bulk)){
|
|
.dma = dma,
|
|
__libeth_xdp_tx_len(skb_frag_size(frag)),
|
|
};
|
|
|
|
return true;
|
|
}
|
|
|
|
/**
|
|
* libeth_xdp_xmit_queue_bulk - internal helper for queueing one XDP xmit frame
|
|
* @bq: XDP Tx bulk to queue the frame to
|
|
* @xdpf: XDP frame to queue
|
|
* @flush_bulk: driver callback to flush the bulk to the HW queue
|
|
*
|
|
* Return: ``LIBETH_XDP_TX`` on success,
|
|
* ``LIBETH_XDP_DROP`` if the frame should be dropped by the stack,
|
|
* ``LIBETH_XDP_ABORTED`` if the frame will be dropped by libeth_xdp.
|
|
*/
|
|
static __always_inline u32
|
|
libeth_xdp_xmit_queue_bulk(struct libeth_xdp_tx_bulk *bq,
|
|
struct xdp_frame *xdpf,
|
|
bool (*flush_bulk)(struct libeth_xdp_tx_bulk *bq,
|
|
u32 flags))
|
|
{
|
|
u32 head, nr_frags, i, ret = LIBETH_XDP_TX;
|
|
struct device *dev = bq->dev->dev.parent;
|
|
const struct skb_shared_info *sinfo;
|
|
|
|
if (unlikely(bq->count == LIBETH_XDP_TX_BULK) &&
|
|
unlikely(!flush_bulk(bq, LIBETH_XDP_TX_NDO)))
|
|
return LIBETH_XDP_DROP;
|
|
|
|
head = libeth_xdp_xmit_queue_head(bq, xdpf, dev);
|
|
if (head == LIBETH_XDP_PASS)
|
|
goto out;
|
|
else if (head == LIBETH_XDP_DROP)
|
|
return LIBETH_XDP_DROP;
|
|
|
|
sinfo = xdp_get_shared_info_from_frame(xdpf);
|
|
nr_frags = sinfo->nr_frags;
|
|
|
|
for (i = 0; i < nr_frags; i++) {
|
|
if (unlikely(bq->count == LIBETH_XDP_TX_BULK) &&
|
|
unlikely(!flush_bulk(bq, LIBETH_XDP_TX_NDO)))
|
|
break;
|
|
|
|
if (!libeth_xdp_xmit_queue_frag(bq, &sinfo->frags[i], dev))
|
|
break;
|
|
}
|
|
|
|
if (unlikely(i < nr_frags))
|
|
ret = LIBETH_XDP_ABORTED;
|
|
|
|
out:
|
|
bq->bulk[bq->count - 1].flags |= LIBETH_XDP_TX_LAST;
|
|
|
|
return ret;
|
|
}
|
|
|
|
/**
|
|
* libeth_xdp_xmit_fill_buf - internal helper to fill one XDP xmit &libeth_sqe
|
|
* @frm: XDP Tx frame from the bulk
|
|
* @i: index on the HW queue
|
|
* @sq: XDPSQ abstraction for the queue
|
|
* @priv: private data
|
|
*
|
|
* Return: XDP Tx descriptor with the mapped DMA and other info to pass to
|
|
* the driver callback.
|
|
*/
|
|
static inline struct libeth_xdp_tx_desc
|
|
libeth_xdp_xmit_fill_buf(struct libeth_xdp_tx_frame frm, u32 i,
|
|
const struct libeth_xdpsq *sq, u64 priv)
|
|
{
|
|
struct libeth_xdp_tx_desc desc;
|
|
struct libeth_sqe *sqe;
|
|
struct xdp_frame *xdpf;
|
|
|
|
if (frm.flags & LIBETH_XDP_TX_FIRST) {
|
|
xdpf = frm.xdpf;
|
|
desc.addr = *libeth_xdp_xmit_frame_dma(xdpf);
|
|
} else {
|
|
xdpf = NULL;
|
|
desc.addr = frm.dma;
|
|
}
|
|
desc.opts = frm.opts;
|
|
|
|
sqe = &sq->sqes[i];
|
|
dma_unmap_addr_set(sqe, dma, desc.addr);
|
|
dma_unmap_len_set(sqe, len, desc.len);
|
|
|
|
if (!xdpf) {
|
|
sqe->type = LIBETH_SQE_XDP_XMIT_FRAG;
|
|
return desc;
|
|
}
|
|
|
|
sqe->type = LIBETH_SQE_XDP_XMIT;
|
|
sqe->xdpf = xdpf;
|
|
libeth_xdp_tx_fill_stats(sqe, &desc,
|
|
xdp_get_shared_info_from_frame(xdpf));
|
|
|
|
return desc;
|
|
}
|
|
|
|
/**
|
|
* libeth_xdp_xmit_flush_bulk - wrapper to define flush of one XDP xmit bulk
|
|
* @bq: bulk to flush
|
|
* @flags: Tx flags, see __libeth_xdp_tx_flush_bulk()
|
|
* @prep: driver callback to prepare the queue
|
|
* @xmit: driver callback to fill a HW descriptor
|
|
*
|
|
* Use via LIBETH_XDP_DEFINE_FLUSH_XMIT() to define an XDP xmit driver
|
|
* callback.
|
|
*/
|
|
#define libeth_xdp_xmit_flush_bulk(bq, flags, prep, xmit) \
|
|
__libeth_xdp_tx_flush_bulk(bq, (flags) | LIBETH_XDP_TX_NDO, prep, \
|
|
libeth_xdp_xmit_fill_buf, xmit)
|
|
|
|
u32 libeth_xdp_xmit_return_bulk(const struct libeth_xdp_tx_frame *bq,
|
|
u32 count, const struct net_device *dev);
|
|
|
|
/**
|
|
* __libeth_xdp_xmit_do_bulk - internal function to implement .ndo_xdp_xmit()
|
|
* @bq: XDP Tx bulk to queue frames to
|
|
* @frames: XDP frames passed by the stack
|
|
* @n: number of frames
|
|
* @flags: flags passed by the stack
|
|
* @flush_bulk: driver callback to flush an XDP xmit bulk
|
|
* @finalize: driver callback to finalize sending XDP Tx frames on the queue
|
|
*
|
|
* Perform common checks, map the frags and queue them to the bulk, then flush
|
|
* the bulk to the XDPSQ. If requested by the stack, finalize the queue.
|
|
*
|
|
* Return: number of frames send or -errno on error.
|
|
*/
|
|
static __always_inline int
|
|
__libeth_xdp_xmit_do_bulk(struct libeth_xdp_tx_bulk *bq,
|
|
struct xdp_frame **frames, u32 n, u32 flags,
|
|
bool (*flush_bulk)(struct libeth_xdp_tx_bulk *bq,
|
|
u32 flags),
|
|
void (*finalize)(void *xdpsq, bool sent, bool flush))
|
|
{
|
|
u32 nxmit = 0;
|
|
|
|
if (unlikely(flags & ~XDP_XMIT_FLAGS_MASK))
|
|
return -EINVAL;
|
|
|
|
for (u32 i = 0; likely(i < n); i++) {
|
|
u32 ret;
|
|
|
|
ret = libeth_xdp_xmit_queue_bulk(bq, frames[i], flush_bulk);
|
|
if (unlikely(ret != LIBETH_XDP_TX)) {
|
|
nxmit += ret == LIBETH_XDP_ABORTED;
|
|
break;
|
|
}
|
|
|
|
nxmit++;
|
|
}
|
|
|
|
if (bq->count) {
|
|
flush_bulk(bq, LIBETH_XDP_TX_NDO);
|
|
if (unlikely(bq->count))
|
|
nxmit -= libeth_xdp_xmit_return_bulk(bq->bulk,
|
|
bq->count,
|
|
bq->dev);
|
|
}
|
|
|
|
finalize(bq->xdpsq, nxmit, flags & XDP_XMIT_FLUSH);
|
|
|
|
return nxmit;
|
|
}
|
|
|
|
/**
|
|
* libeth_xdp_xmit_do_bulk - implement full .ndo_xdp_xmit() in driver
|
|
* @dev: target &net_device
|
|
* @n: number of frames to send
|
|
* @fr: XDP frames to send
|
|
* @f: flags passed by the stack
|
|
* @xqs: array of XDPSQs driver structs
|
|
* @nqs: number of active XDPSQs, the above array length
|
|
* @fl: driver callback to flush an XDP xmit bulk
|
|
* @fin: driver cabback to finalize the queue
|
|
*
|
|
* If the driver has active XDPSQs, perform common checks and send the frames.
|
|
* Finalize the queue, if requested.
|
|
*
|
|
* Return: number of frames sent or -errno on error.
|
|
*/
|
|
#define libeth_xdp_xmit_do_bulk(dev, n, fr, f, xqs, nqs, fl, fin) \
|
|
_libeth_xdp_xmit_do_bulk(dev, n, fr, f, xqs, nqs, fl, fin, \
|
|
__UNIQUE_ID(bq_), __UNIQUE_ID(ret_), \
|
|
__UNIQUE_ID(nqs_))
|
|
|
|
#define _libeth_xdp_xmit_do_bulk(d, n, fr, f, xqs, nqs, fl, fin, ub, ur, un) \
|
|
({ \
|
|
u32 un = (nqs); \
|
|
int ur; \
|
|
\
|
|
if (likely(un)) { \
|
|
LIBETH_XDP_ONSTACK_BULK(ub); \
|
|
\
|
|
libeth_xdp_xmit_init_bulk(&ub, d, xqs, un); \
|
|
ur = __libeth_xdp_xmit_do_bulk(&ub, fr, n, f, fl, fin); \
|
|
} else { \
|
|
ur = -ENXIO; \
|
|
} \
|
|
\
|
|
ur; \
|
|
})
|
|
|
|
/* Rx polling path */
|
|
|
|
/**
|
|
* libeth_xdp_tx_init_bulk - initialize an XDP Tx bulk for Rx NAPI poll
|
|
* @bq: bulk to initialize
|
|
* @prog: RCU pointer to the XDP program (can be %NULL)
|
|
* @dev: target &net_device
|
|
* @xdpsqs: array of driver XDPSQ structs
|
|
* @num: number of active XDPSQs, the above array length
|
|
*
|
|
* Should be called on an onstack XDP Tx bulk before the NAPI polling loop.
|
|
* Initializes all the needed fields to run libeth_xdp functions. If @num == 0,
|
|
* assumes XDP is not enabled.
|
|
* Do not use for XSk, it has its own optimized helper.
|
|
*/
|
|
#define libeth_xdp_tx_init_bulk(bq, prog, dev, xdpsqs, num) \
|
|
__libeth_xdp_tx_init_bulk(bq, prog, dev, xdpsqs, num, false, \
|
|
__UNIQUE_ID(bq_), __UNIQUE_ID(nqs_))
|
|
|
|
#define __libeth_xdp_tx_init_bulk(bq, pr, d, xdpsqs, num, xsk, ub, un) do { \
|
|
typeof(bq) ub = (bq); \
|
|
u32 un = (num); \
|
|
\
|
|
rcu_read_lock(); \
|
|
\
|
|
if (un || (xsk)) { \
|
|
ub->prog = rcu_dereference(pr); \
|
|
ub->dev = (d); \
|
|
ub->xdpsq = (xdpsqs)[libeth_xdpsq_id(un)]; \
|
|
} else { \
|
|
ub->prog = NULL; \
|
|
} \
|
|
\
|
|
ub->act_mask = 0; \
|
|
ub->count = 0; \
|
|
} while (0)
|
|
|
|
void libeth_xdp_load_stash(struct libeth_xdp_buff *dst,
|
|
const struct libeth_xdp_buff_stash *src);
|
|
void libeth_xdp_save_stash(struct libeth_xdp_buff_stash *dst,
|
|
const struct libeth_xdp_buff *src);
|
|
void __libeth_xdp_return_stash(struct libeth_xdp_buff_stash *stash);
|
|
|
|
/**
|
|
* libeth_xdp_init_buff - initialize a &libeth_xdp_buff for Rx NAPI poll
|
|
* @dst: onstack buffer to initialize
|
|
* @src: XDP buffer stash placed on the queue
|
|
* @rxq: registered &xdp_rxq_info corresponding to this queue
|
|
*
|
|
* Should be called before the main NAPI polling loop. Loads the content of
|
|
* the previously saved stash or initializes the buffer from scratch.
|
|
* Do not use for XSk.
|
|
*/
|
|
static inline void
|
|
libeth_xdp_init_buff(struct libeth_xdp_buff *dst,
|
|
const struct libeth_xdp_buff_stash *src,
|
|
struct xdp_rxq_info *rxq)
|
|
{
|
|
if (likely(!src->data))
|
|
dst->data = NULL;
|
|
else
|
|
libeth_xdp_load_stash(dst, src);
|
|
|
|
dst->base.rxq = rxq;
|
|
}
|
|
|
|
/**
|
|
* libeth_xdp_save_buff - save a partially built buffer on a queue
|
|
* @dst: XDP buffer stash placed on the queue
|
|
* @src: onstack buffer to save
|
|
*
|
|
* Should be called after the main NAPI polling loop. If the loop exited before
|
|
* the buffer was finished, saves its content on the queue, so that it can be
|
|
* completed during the next poll. Otherwise, clears the stash.
|
|
*/
|
|
static inline void libeth_xdp_save_buff(struct libeth_xdp_buff_stash *dst,
|
|
const struct libeth_xdp_buff *src)
|
|
{
|
|
if (likely(!src->data))
|
|
dst->data = NULL;
|
|
else
|
|
libeth_xdp_save_stash(dst, src);
|
|
}
|
|
|
|
/**
|
|
* libeth_xdp_return_stash - free an XDP buffer stash from a queue
|
|
* @stash: stash to free
|
|
*
|
|
* If the queue is about to be destroyed, but it still has an incompleted
|
|
* buffer stash, this helper should be called to free it.
|
|
*/
|
|
static inline void libeth_xdp_return_stash(struct libeth_xdp_buff_stash *stash)
|
|
{
|
|
if (stash->data)
|
|
__libeth_xdp_return_stash(stash);
|
|
}
|
|
|
|
static inline void libeth_xdp_return_va(const void *data, bool napi)
|
|
{
|
|
netmem_ref netmem = virt_to_netmem(data);
|
|
|
|
page_pool_put_full_netmem(__netmem_get_pp(netmem), netmem, napi);
|
|
}
|
|
|
|
static inline void libeth_xdp_return_frags(const struct skb_shared_info *sinfo,
|
|
bool napi)
|
|
{
|
|
for (u32 i = 0; i < sinfo->nr_frags; i++) {
|
|
netmem_ref netmem = skb_frag_netmem(&sinfo->frags[i]);
|
|
|
|
page_pool_put_full_netmem(netmem_get_pp(netmem), netmem, napi);
|
|
}
|
|
}
|
|
|
|
/**
|
|
* libeth_xdp_return_buff - free/recycle &libeth_xdp_buff
|
|
* @xdp: buffer to free
|
|
*
|
|
* Hotpath helper to free &libeth_xdp_buff. Comparing to xdp_return_buff(),
|
|
* it's faster as it gets inlined and always assumes order-0 pages and safe
|
|
* direct recycling. Zeroes @xdp->data to avoid UAFs.
|
|
*/
|
|
#define libeth_xdp_return_buff(xdp) __libeth_xdp_return_buff(xdp, true)
|
|
|
|
static inline void __libeth_xdp_return_buff(struct libeth_xdp_buff *xdp,
|
|
bool napi)
|
|
{
|
|
if (!xdp_buff_has_frags(&xdp->base))
|
|
goto out;
|
|
|
|
libeth_xdp_return_frags(xdp_get_shared_info_from_buff(&xdp->base),
|
|
napi);
|
|
|
|
out:
|
|
libeth_xdp_return_va(xdp->data, napi);
|
|
xdp->data = NULL;
|
|
}
|
|
|
|
bool libeth_xdp_buff_add_frag(struct libeth_xdp_buff *xdp,
|
|
const struct libeth_fqe *fqe,
|
|
u32 len);
|
|
|
|
/**
|
|
* libeth_xdp_prepare_buff - fill &libeth_xdp_buff with head FQE data
|
|
* @xdp: XDP buffer to attach the head to
|
|
* @fqe: FQE containing the head buffer
|
|
* @len: buffer len passed from HW
|
|
*
|
|
* Internal, use libeth_xdp_process_buff() instead. Initializes XDP buffer
|
|
* head with the Rx buffer data: data pointer, length, headroom, and
|
|
* truesize/tailroom. Zeroes the flags.
|
|
* Uses faster single u64 write instead of per-field access.
|
|
*/
|
|
static inline void libeth_xdp_prepare_buff(struct libeth_xdp_buff *xdp,
|
|
const struct libeth_fqe *fqe,
|
|
u32 len)
|
|
{
|
|
const struct page *page = __netmem_to_page(fqe->netmem);
|
|
|
|
#ifdef __LIBETH_WORD_ACCESS
|
|
static_assert(offsetofend(typeof(xdp->base), flags) -
|
|
offsetof(typeof(xdp->base), frame_sz) ==
|
|
sizeof(u64));
|
|
|
|
*(u64 *)&xdp->base.frame_sz = fqe->truesize;
|
|
#else
|
|
xdp_init_buff(&xdp->base, fqe->truesize, xdp->base.rxq);
|
|
#endif
|
|
xdp_prepare_buff(&xdp->base, page_address(page) + fqe->offset,
|
|
pp_page_to_nmdesc(page)->pp->p.offset, len, true);
|
|
}
|
|
|
|
/**
|
|
* libeth_xdp_process_buff - attach Rx buffer to &libeth_xdp_buff
|
|
* @xdp: XDP buffer to attach the Rx buffer to
|
|
* @fqe: Rx buffer to process
|
|
* @len: received data length from the descriptor
|
|
*
|
|
* If the XDP buffer is empty, attaches the Rx buffer as head and initializes
|
|
* the required fields. Otherwise, attaches the buffer as a frag.
|
|
* Already performs DMA sync-for-CPU and frame start prefetch
|
|
* (for head buffers only).
|
|
*
|
|
* Return: true on success, false if the descriptor must be skipped (empty or
|
|
* no space for a new frag).
|
|
*/
|
|
static inline bool libeth_xdp_process_buff(struct libeth_xdp_buff *xdp,
|
|
const struct libeth_fqe *fqe,
|
|
u32 len)
|
|
{
|
|
if (!libeth_rx_sync_for_cpu(fqe, len))
|
|
return false;
|
|
|
|
if (xdp->data)
|
|
return libeth_xdp_buff_add_frag(xdp, fqe, len);
|
|
|
|
libeth_xdp_prepare_buff(xdp, fqe, len);
|
|
|
|
prefetch(xdp->data);
|
|
|
|
return true;
|
|
}
|
|
|
|
/**
|
|
* libeth_xdp_buff_stats_frags - update onstack RQ stats with XDP frags info
|
|
* @ss: onstack stats to update
|
|
* @xdp: buffer to account
|
|
*
|
|
* Internal helper used by __libeth_xdp_run_pass(), do not call directly.
|
|
* Adds buffer's frags count and total len to the onstack stats.
|
|
*/
|
|
static inline void
|
|
libeth_xdp_buff_stats_frags(struct libeth_rq_napi_stats *ss,
|
|
const struct libeth_xdp_buff *xdp)
|
|
{
|
|
const struct skb_shared_info *sinfo;
|
|
|
|
sinfo = xdp_get_shared_info_from_buff(&xdp->base);
|
|
ss->bytes += sinfo->xdp_frags_size;
|
|
ss->fragments += sinfo->nr_frags + 1;
|
|
}
|
|
|
|
u32 libeth_xdp_prog_exception(const struct libeth_xdp_tx_bulk *bq,
|
|
struct libeth_xdp_buff *xdp,
|
|
enum xdp_action act, int ret);
|
|
|
|
/**
|
|
* __libeth_xdp_run_prog - run XDP program on an XDP buffer
|
|
* @xdp: XDP buffer to run the prog on
|
|
* @bq: buffer bulk for ``XDP_TX`` queueing
|
|
*
|
|
* Internal inline abstraction to run XDP program. Handles ``XDP_DROP``
|
|
* and ``XDP_REDIRECT`` only, the rest is processed levels up.
|
|
* Reports an XDP prog exception on errors.
|
|
*
|
|
* Return: libeth_xdp prog verdict depending on the prog's verdict.
|
|
*/
|
|
static __always_inline u32
|
|
__libeth_xdp_run_prog(struct libeth_xdp_buff *xdp,
|
|
const struct libeth_xdp_tx_bulk *bq)
|
|
{
|
|
enum xdp_action act;
|
|
|
|
act = bpf_prog_run_xdp(bq->prog, &xdp->base);
|
|
if (unlikely(act < XDP_DROP || act > XDP_REDIRECT))
|
|
goto out;
|
|
|
|
switch (act) {
|
|
case XDP_PASS:
|
|
return LIBETH_XDP_PASS;
|
|
case XDP_DROP:
|
|
libeth_xdp_return_buff(xdp);
|
|
|
|
return LIBETH_XDP_DROP;
|
|
case XDP_TX:
|
|
return LIBETH_XDP_TX;
|
|
case XDP_REDIRECT:
|
|
if (unlikely(xdp_do_redirect(bq->dev, &xdp->base, bq->prog)))
|
|
break;
|
|
|
|
xdp->data = NULL;
|
|
|
|
return LIBETH_XDP_REDIRECT;
|
|
default:
|
|
break;
|
|
}
|
|
|
|
out:
|
|
return libeth_xdp_prog_exception(bq, xdp, act, 0);
|
|
}
|
|
|
|
/**
|
|
* __libeth_xdp_run_flush - run XDP program and handle ``XDP_TX`` verdict
|
|
* @xdp: XDP buffer to run the prog on
|
|
* @bq: buffer bulk for ``XDP_TX`` queueing
|
|
* @run: internal callback for running XDP program
|
|
* @queue: internal callback for queuing ``XDP_TX`` frame
|
|
* @flush_bulk: driver callback for flushing a bulk
|
|
*
|
|
* Internal inline abstraction to run XDP program and additionally handle
|
|
* ``XDP_TX`` verdict. Used by both XDP and XSk, hence @run and @queue.
|
|
* Do not use directly.
|
|
*
|
|
* Return: libeth_xdp prog verdict depending on the prog's verdict.
|
|
*/
|
|
static __always_inline u32
|
|
__libeth_xdp_run_flush(struct libeth_xdp_buff *xdp,
|
|
struct libeth_xdp_tx_bulk *bq,
|
|
u32 (*run)(struct libeth_xdp_buff *xdp,
|
|
const struct libeth_xdp_tx_bulk *bq),
|
|
bool (*queue)(struct libeth_xdp_tx_bulk *bq,
|
|
struct libeth_xdp_buff *xdp,
|
|
bool (*flush_bulk)
|
|
(struct libeth_xdp_tx_bulk *bq,
|
|
u32 flags)),
|
|
bool (*flush_bulk)(struct libeth_xdp_tx_bulk *bq,
|
|
u32 flags))
|
|
{
|
|
u32 act;
|
|
|
|
act = run(xdp, bq);
|
|
if (act == LIBETH_XDP_TX && unlikely(!queue(bq, xdp, flush_bulk)))
|
|
act = LIBETH_XDP_DROP;
|
|
|
|
bq->act_mask |= act;
|
|
|
|
return act;
|
|
}
|
|
|
|
/**
|
|
* libeth_xdp_run_prog - run XDP program (non-XSk path) and handle all verdicts
|
|
* @xdp: XDP buffer to process
|
|
* @bq: XDP Tx bulk to queue ``XDP_TX`` buffers
|
|
* @fl: driver ``XDP_TX`` bulk flush callback
|
|
*
|
|
* Run the attached XDP program and handle all possible verdicts. XSk has its
|
|
* own version.
|
|
* Prefer using it via LIBETH_XDP_DEFINE_RUN{,_PASS,_PROG}().
|
|
*
|
|
* Return: true if the buffer should be passed up the stack, false if the poll
|
|
* should go to the next buffer.
|
|
*/
|
|
#define libeth_xdp_run_prog(xdp, bq, fl) \
|
|
(__libeth_xdp_run_flush(xdp, bq, __libeth_xdp_run_prog, \
|
|
libeth_xdp_tx_queue_bulk, \
|
|
fl) == LIBETH_XDP_PASS)
|
|
|
|
/**
|
|
* __libeth_xdp_run_pass - helper to run XDP program and handle the result
|
|
* @xdp: XDP buffer to process
|
|
* @bq: XDP Tx bulk to queue ``XDP_TX`` frames
|
|
* @napi: NAPI to build an skb and pass it up the stack
|
|
* @rs: onstack libeth RQ stats
|
|
* @md: metadata that should be filled to the XDP buffer
|
|
* @prep: callback for filling the metadata
|
|
* @run: driver wrapper to run XDP program
|
|
* @populate: driver callback to populate an skb with the HW descriptor data
|
|
*
|
|
* Inline abstraction that does the following (non-XSk path):
|
|
* 1) adds frame size and frag number (if needed) to the onstack stats;
|
|
* 2) fills the descriptor metadata to the onstack &libeth_xdp_buff
|
|
* 3) runs XDP program if present;
|
|
* 4) handles all possible verdicts;
|
|
* 5) on ``XDP_PASS`, builds an skb from the buffer;
|
|
* 6) populates it with the descriptor metadata;
|
|
* 7) passes it up the stack.
|
|
*
|
|
* In most cases, number 2 means just writing the pointer to the HW descriptor
|
|
* to the XDP buffer. If so, please use LIBETH_XDP_DEFINE_RUN{,_PASS}()
|
|
* wrappers to build a driver function.
|
|
*/
|
|
static __always_inline void
|
|
__libeth_xdp_run_pass(struct libeth_xdp_buff *xdp,
|
|
struct libeth_xdp_tx_bulk *bq, struct napi_struct *napi,
|
|
struct libeth_rq_napi_stats *rs, const void *md,
|
|
void (*prep)(struct libeth_xdp_buff *xdp,
|
|
const void *md),
|
|
bool (*run)(struct libeth_xdp_buff *xdp,
|
|
struct libeth_xdp_tx_bulk *bq),
|
|
bool (*populate)(struct sk_buff *skb,
|
|
const struct libeth_xdp_buff *xdp,
|
|
struct libeth_rq_napi_stats *rs))
|
|
{
|
|
struct sk_buff *skb;
|
|
|
|
rs->bytes += xdp->base.data_end - xdp->data;
|
|
rs->packets++;
|
|
|
|
if (xdp_buff_has_frags(&xdp->base))
|
|
libeth_xdp_buff_stats_frags(rs, xdp);
|
|
|
|
if (prep && (!__builtin_constant_p(!!md) || md))
|
|
prep(xdp, md);
|
|
|
|
if (!bq || !run || !bq->prog)
|
|
goto build;
|
|
|
|
if (!run(xdp, bq))
|
|
return;
|
|
|
|
build:
|
|
skb = xdp_build_skb_from_buff(&xdp->base);
|
|
if (unlikely(!skb)) {
|
|
libeth_xdp_return_buff_slow(xdp);
|
|
return;
|
|
}
|
|
|
|
xdp->data = NULL;
|
|
|
|
if (unlikely(!populate(skb, xdp, rs))) {
|
|
napi_consume_skb(skb, true);
|
|
return;
|
|
}
|
|
|
|
napi_gro_receive(napi, skb);
|
|
}
|
|
|
|
static inline void libeth_xdp_prep_desc(struct libeth_xdp_buff *xdp,
|
|
const void *desc)
|
|
{
|
|
xdp->desc = desc;
|
|
}
|
|
|
|
/**
|
|
* libeth_xdp_run_pass - helper to run XDP program and handle the result
|
|
* @xdp: XDP buffer to process
|
|
* @bq: XDP Tx bulk to queue ``XDP_TX`` frames
|
|
* @napi: NAPI to build an skb and pass it up the stack
|
|
* @ss: onstack libeth RQ stats
|
|
* @desc: pointer to the HW descriptor for that frame
|
|
* @run: driver wrapper to run XDP program
|
|
* @populate: driver callback to populate an skb with the HW descriptor data
|
|
*
|
|
* Wrapper around the underscored version when "fill the descriptor metadata"
|
|
* means just writing the pointer to the HW descriptor as @xdp->desc.
|
|
*/
|
|
#define libeth_xdp_run_pass(xdp, bq, napi, ss, desc, run, populate) \
|
|
__libeth_xdp_run_pass(xdp, bq, napi, ss, desc, libeth_xdp_prep_desc, \
|
|
run, populate)
|
|
|
|
/**
|
|
* libeth_xdp_finalize_rx - finalize XDPSQ after a NAPI polling loop (non-XSk)
|
|
* @bq: ``XDP_TX`` frame bulk
|
|
* @flush: driver callback to flush the bulk
|
|
* @finalize: driver callback to start sending the frames and run the timer
|
|
*
|
|
* Flush the bulk if there are frames left to send, kick the queue and flush
|
|
* the XDP maps.
|
|
*/
|
|
#define libeth_xdp_finalize_rx(bq, flush, finalize) \
|
|
__libeth_xdp_finalize_rx(bq, 0, flush, finalize)
|
|
|
|
static __always_inline void
|
|
__libeth_xdp_finalize_rx(struct libeth_xdp_tx_bulk *bq, u32 flags,
|
|
bool (*flush_bulk)(struct libeth_xdp_tx_bulk *bq,
|
|
u32 flags),
|
|
void (*finalize)(void *xdpsq, bool sent, bool flush))
|
|
{
|
|
if (bq->act_mask & LIBETH_XDP_TX) {
|
|
if (bq->count)
|
|
flush_bulk(bq, flags | LIBETH_XDP_TX_DROP);
|
|
finalize(bq->xdpsq, true, true);
|
|
}
|
|
if (bq->act_mask & LIBETH_XDP_REDIRECT)
|
|
xdp_do_flush();
|
|
|
|
rcu_read_unlock();
|
|
}
|
|
|
|
/*
|
|
* Helpers to reduce boilerplate code in drivers.
|
|
*
|
|
* Typical driver Rx flow would be (excl. bulk and buff init, frag attach):
|
|
*
|
|
* LIBETH_XDP_DEFINE_START();
|
|
* LIBETH_XDP_DEFINE_FLUSH_TX(static driver_xdp_flush_tx, driver_xdp_tx_prep,
|
|
* driver_xdp_xmit);
|
|
* LIBETH_XDP_DEFINE_RUN(static driver_xdp_run, driver_xdp_run_prog,
|
|
* driver_xdp_flush_tx, driver_populate_skb);
|
|
* LIBETH_XDP_DEFINE_FINALIZE(static driver_xdp_finalize_rx,
|
|
* driver_xdp_flush_tx, driver_xdp_finalize_sq);
|
|
* LIBETH_XDP_DEFINE_END();
|
|
*
|
|
* This will build a set of 4 static functions. The compiler is free to decide
|
|
* whether to inline them.
|
|
* Then, in the NAPI polling function:
|
|
*
|
|
* while (packets < budget) {
|
|
* // ...
|
|
* driver_xdp_run(xdp, &bq, napi, &rs, desc);
|
|
* }
|
|
* driver_xdp_finalize_rx(&bq);
|
|
*/
|
|
|
|
#define LIBETH_XDP_DEFINE_START() \
|
|
__diag_push(); \
|
|
__diag_ignore(GCC, 8, "-Wold-style-declaration", \
|
|
"Allow specifying \'static\' after the return type")
|
|
|
|
/**
|
|
* LIBETH_XDP_DEFINE_TIMER - define a driver XDPSQ cleanup timer callback
|
|
* @name: name of the function to define
|
|
* @poll: Tx polling/completion function
|
|
*/
|
|
#define LIBETH_XDP_DEFINE_TIMER(name, poll) \
|
|
void name(struct work_struct *work) \
|
|
{ \
|
|
libeth_xdpsq_run_timer(work, poll); \
|
|
}
|
|
|
|
/**
|
|
* LIBETH_XDP_DEFINE_FLUSH_TX - define a driver ``XDP_TX`` bulk flush function
|
|
* @name: name of the function to define
|
|
* @prep: driver callback to clean an XDPSQ
|
|
* @xmit: driver callback to write a HW Tx descriptor
|
|
*/
|
|
#define LIBETH_XDP_DEFINE_FLUSH_TX(name, prep, xmit) \
|
|
__LIBETH_XDP_DEFINE_FLUSH_TX(name, prep, xmit, xdp)
|
|
|
|
#define __LIBETH_XDP_DEFINE_FLUSH_TX(name, prep, xmit, pfx) \
|
|
bool name(struct libeth_xdp_tx_bulk *bq, u32 flags) \
|
|
{ \
|
|
return libeth_##pfx##_tx_flush_bulk(bq, flags, prep, xmit); \
|
|
}
|
|
|
|
/**
|
|
* LIBETH_XDP_DEFINE_FLUSH_XMIT - define a driver XDP xmit bulk flush function
|
|
* @name: name of the function to define
|
|
* @prep: driver callback to clean an XDPSQ
|
|
* @xmit: driver callback to write a HW Tx descriptor
|
|
*/
|
|
#define LIBETH_XDP_DEFINE_FLUSH_XMIT(name, prep, xmit) \
|
|
bool name(struct libeth_xdp_tx_bulk *bq, u32 flags) \
|
|
{ \
|
|
return libeth_xdp_xmit_flush_bulk(bq, flags, prep, xmit); \
|
|
}
|
|
|
|
/**
|
|
* LIBETH_XDP_DEFINE_RUN_PROG - define a driver XDP program run function
|
|
* @name: name of the function to define
|
|
* @flush: driver callback to flush an ``XDP_TX`` bulk
|
|
*/
|
|
#define LIBETH_XDP_DEFINE_RUN_PROG(name, flush) \
|
|
bool __LIBETH_XDP_DEFINE_RUN_PROG(name, flush, xdp)
|
|
|
|
#define __LIBETH_XDP_DEFINE_RUN_PROG(name, flush, pfx) \
|
|
name(struct libeth_xdp_buff *xdp, struct libeth_xdp_tx_bulk *bq) \
|
|
{ \
|
|
return libeth_##pfx##_run_prog(xdp, bq, flush); \
|
|
}
|
|
|
|
/**
|
|
* LIBETH_XDP_DEFINE_RUN_PASS - define a driver buffer process + pass function
|
|
* @name: name of the function to define
|
|
* @run: driver callback to run XDP program (above)
|
|
* @populate: driver callback to fill an skb with HW descriptor info
|
|
*/
|
|
#define LIBETH_XDP_DEFINE_RUN_PASS(name, run, populate) \
|
|
void __LIBETH_XDP_DEFINE_RUN_PASS(name, run, populate, xdp)
|
|
|
|
#define __LIBETH_XDP_DEFINE_RUN_PASS(name, run, populate, pfx) \
|
|
name(struct libeth_xdp_buff *xdp, struct libeth_xdp_tx_bulk *bq, \
|
|
struct napi_struct *napi, struct libeth_rq_napi_stats *ss, \
|
|
const void *desc) \
|
|
{ \
|
|
return libeth_##pfx##_run_pass(xdp, bq, napi, ss, desc, run, \
|
|
populate); \
|
|
}
|
|
|
|
/**
|
|
* LIBETH_XDP_DEFINE_RUN - define a driver buffer process, run + pass function
|
|
* @name: name of the function to define
|
|
* @run: name of the XDP prog run function to define
|
|
* @flush: driver callback to flush an ``XDP_TX`` bulk
|
|
* @populate: driver callback to fill an skb with HW descriptor info
|
|
*/
|
|
#define LIBETH_XDP_DEFINE_RUN(name, run, flush, populate) \
|
|
__LIBETH_XDP_DEFINE_RUN(name, run, flush, populate, XDP)
|
|
|
|
#define __LIBETH_XDP_DEFINE_RUN(name, run, flush, populate, pfx) \
|
|
LIBETH_##pfx##_DEFINE_RUN_PROG(static run, flush); \
|
|
LIBETH_##pfx##_DEFINE_RUN_PASS(name, run, populate)
|
|
|
|
/**
|
|
* LIBETH_XDP_DEFINE_FINALIZE - define a driver Rx NAPI poll finalize function
|
|
* @name: name of the function to define
|
|
* @flush: driver callback to flush an ``XDP_TX`` bulk
|
|
* @finalize: driver callback to finalize an XDPSQ and run the timer
|
|
*/
|
|
#define LIBETH_XDP_DEFINE_FINALIZE(name, flush, finalize) \
|
|
__LIBETH_XDP_DEFINE_FINALIZE(name, flush, finalize, xdp)
|
|
|
|
#define __LIBETH_XDP_DEFINE_FINALIZE(name, flush, finalize, pfx) \
|
|
void name(struct libeth_xdp_tx_bulk *bq) \
|
|
{ \
|
|
libeth_##pfx##_finalize_rx(bq, flush, finalize); \
|
|
}
|
|
|
|
#define LIBETH_XDP_DEFINE_END() __diag_pop()
|
|
|
|
/* XMO */
|
|
|
|
/**
|
|
* libeth_xdp_buff_to_rq - get RQ pointer from an XDP buffer pointer
|
|
* @xdp: &libeth_xdp_buff corresponding to the queue
|
|
* @type: typeof() of the driver Rx queue structure
|
|
* @member: name of &xdp_rxq_info inside @type
|
|
*
|
|
* Often times, pointer to the RQ is needed when reading/filling metadata from
|
|
* HW descriptors. The helper can be used to quickly jump from an XDP buffer
|
|
* to the queue corresponding to its &xdp_rxq_info without introducing
|
|
* additional fields (&libeth_xdp_buff is precisely 1 cacheline long on x64).
|
|
*/
|
|
#define libeth_xdp_buff_to_rq(xdp, type, member) \
|
|
container_of_const((xdp)->base.rxq, type, member)
|
|
|
|
/**
|
|
* libeth_xdpmo_rx_hash - convert &libeth_rx_pt to an XDP RSS hash metadata
|
|
* @hash: pointer to the variable to write the hash to
|
|
* @rss_type: pointer to the variable to write the hash type to
|
|
* @val: hash value from the HW descriptor
|
|
* @pt: libeth parsed packet type
|
|
*
|
|
* Handle zeroed/non-available hash and convert libeth parsed packet type to
|
|
* the corresponding XDP RSS hash type. To be called at the end of
|
|
* xdp_metadata_ops idpf_xdpmo::xmo_rx_hash() implementation.
|
|
* Note that if the driver doesn't use a constant packet type lookup table but
|
|
* generates it at runtime, it must call libeth_rx_pt_gen_hash_type(pt) to
|
|
* generate XDP RSS hash type for each packet type.
|
|
*
|
|
* Return: 0 on success, -ENODATA when the hash is not available.
|
|
*/
|
|
static inline int libeth_xdpmo_rx_hash(u32 *hash,
|
|
enum xdp_rss_hash_type *rss_type,
|
|
u32 val, struct libeth_rx_pt pt)
|
|
{
|
|
if (unlikely(!val))
|
|
return -ENODATA;
|
|
|
|
*hash = val;
|
|
*rss_type = pt.hash_type;
|
|
|
|
return 0;
|
|
}
|
|
|
|
/* Tx buffer completion */
|
|
|
|
void libeth_xdp_return_buff_bulk(const struct skb_shared_info *sinfo,
|
|
struct xdp_frame_bulk *bq, bool frags);
|
|
void libeth_xsk_buff_free_slow(struct libeth_xdp_buff *xdp);
|
|
|
|
/**
|
|
* __libeth_xdp_complete_tx - complete sent XDPSQE
|
|
* @sqe: SQ element / Tx buffer to complete
|
|
* @cp: Tx polling/completion params
|
|
* @bulk: internal callback to bulk-free ``XDP_TX`` buffers
|
|
* @xsk: internal callback to free XSk ``XDP_TX`` buffers
|
|
*
|
|
* Use the non-underscored version in drivers instead. This one is shared
|
|
* internally with libeth_tx_complete_any().
|
|
* Complete an XDPSQE of any type of XDP frame. This includes DMA unmapping
|
|
* when needed, buffer freeing, stats update, and SQE invalidation.
|
|
*/
|
|
static __always_inline void
|
|
__libeth_xdp_complete_tx(struct libeth_sqe *sqe, struct libeth_cq_pp *cp,
|
|
typeof(libeth_xdp_return_buff_bulk) bulk,
|
|
typeof(libeth_xsk_buff_free_slow) xsk)
|
|
{
|
|
enum libeth_sqe_type type = sqe->type;
|
|
|
|
switch (type) {
|
|
case LIBETH_SQE_EMPTY:
|
|
return;
|
|
case LIBETH_SQE_XDP_XMIT:
|
|
case LIBETH_SQE_XDP_XMIT_FRAG:
|
|
dma_unmap_page(cp->dev, dma_unmap_addr(sqe, dma),
|
|
dma_unmap_len(sqe, len), DMA_TO_DEVICE);
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
|
|
switch (type) {
|
|
case LIBETH_SQE_XDP_TX:
|
|
bulk(sqe->sinfo, cp->bq, sqe->nr_frags != 1);
|
|
break;
|
|
case LIBETH_SQE_XDP_XMIT:
|
|
xdp_return_frame_bulk(sqe->xdpf, cp->bq);
|
|
break;
|
|
case LIBETH_SQE_XSK_TX:
|
|
case LIBETH_SQE_XSK_TX_FRAG:
|
|
xsk(sqe->xsk);
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
|
|
switch (type) {
|
|
case LIBETH_SQE_XDP_TX:
|
|
case LIBETH_SQE_XDP_XMIT:
|
|
case LIBETH_SQE_XSK_TX:
|
|
cp->xdp_tx -= sqe->nr_frags;
|
|
|
|
cp->xss->packets++;
|
|
cp->xss->bytes += sqe->bytes;
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
|
|
sqe->type = LIBETH_SQE_EMPTY;
|
|
}
|
|
|
|
static inline void libeth_xdp_complete_tx(struct libeth_sqe *sqe,
|
|
struct libeth_cq_pp *cp)
|
|
{
|
|
__libeth_xdp_complete_tx(sqe, cp, libeth_xdp_return_buff_bulk,
|
|
libeth_xsk_buff_free_slow);
|
|
}
|
|
|
|
/* Misc */
|
|
|
|
u32 libeth_xdp_queue_threshold(u32 count);
|
|
|
|
void __libeth_xdp_set_features(struct net_device *dev,
|
|
const struct xdp_metadata_ops *xmo,
|
|
u32 zc_segs,
|
|
const struct xsk_tx_metadata_ops *tmo);
|
|
void libeth_xdp_set_redirect(struct net_device *dev, bool enable);
|
|
|
|
/**
|
|
* libeth_xdp_set_features - set XDP features for netdev
|
|
* @dev: &net_device to configure
|
|
* @...: optional params, see __libeth_xdp_set_features()
|
|
*
|
|
* Set all the features libeth_xdp supports, including .ndo_xdp_xmit(). That
|
|
* said, it should be used only when XDPSQs are always available regardless
|
|
* of whether an XDP prog is attached to @dev.
|
|
*/
|
|
#define libeth_xdp_set_features(dev, ...) \
|
|
CONCATENATE(__libeth_xdp_feat, \
|
|
COUNT_ARGS(__VA_ARGS__))(dev, ##__VA_ARGS__)
|
|
|
|
#define __libeth_xdp_feat0(dev) \
|
|
__libeth_xdp_set_features(dev, NULL, 0, NULL)
|
|
#define __libeth_xdp_feat1(dev, xmo) \
|
|
__libeth_xdp_set_features(dev, xmo, 0, NULL)
|
|
#define __libeth_xdp_feat2(dev, xmo, zc_segs) \
|
|
__libeth_xdp_set_features(dev, xmo, zc_segs, NULL)
|
|
#define __libeth_xdp_feat3(dev, xmo, zc_segs, tmo) \
|
|
__libeth_xdp_set_features(dev, xmo, zc_segs, tmo)
|
|
|
|
/**
|
|
* libeth_xdp_set_features_noredir - enable all libeth_xdp features w/o redir
|
|
* @dev: target &net_device
|
|
* @...: optional params, see __libeth_xdp_set_features()
|
|
*
|
|
* Enable everything except the .ndo_xdp_xmit() feature, use when XDPSQs are
|
|
* not available right after netdev registration.
|
|
*/
|
|
#define libeth_xdp_set_features_noredir(dev, ...) \
|
|
__libeth_xdp_set_features_noredir(dev, __UNIQUE_ID(dev_), \
|
|
##__VA_ARGS__)
|
|
|
|
#define __libeth_xdp_set_features_noredir(dev, ud, ...) do { \
|
|
struct net_device *ud = (dev); \
|
|
\
|
|
libeth_xdp_set_features(ud, ##__VA_ARGS__); \
|
|
libeth_xdp_set_redirect(ud, false); \
|
|
} while (0)
|
|
|
|
#define libeth_xsktmo ((const void *)GOLDEN_RATIO_PRIME)
|
|
|
|
#endif /* __LIBETH_XDP_H */
|