Lines Matching refs:xdp
65 return pool->heads[0].xdp.rxq->napi_id; in xsk_pool_get_napi_id()
85 static inline dma_addr_t xsk_buff_xdp_get_dma(struct xdp_buff *xdp) in xsk_buff_xdp_get_dma() argument
87 struct xdp_buff_xsk *xskb = container_of(xdp, struct xdp_buff_xsk, xdp); in xsk_buff_xdp_get_dma()
92 static inline dma_addr_t xsk_buff_xdp_get_frame_dma(struct xdp_buff *xdp) in xsk_buff_xdp_get_frame_dma() argument
94 struct xdp_buff_xsk *xskb = container_of(xdp, struct xdp_buff_xsk, xdp); in xsk_buff_xdp_get_frame_dma()
110 static inline u32 xsk_buff_alloc_batch(struct xsk_buff_pool *pool, struct xdp_buff **xdp, u32 max) in xsk_buff_alloc_batch() argument
112 return xp_alloc_batch(pool, xdp, max); in xsk_buff_alloc_batch()
120 static inline void xsk_buff_free(struct xdp_buff *xdp) in xsk_buff_free() argument
122 struct xdp_buff_xsk *xskb = container_of(xdp, struct xdp_buff_xsk, xdp); in xsk_buff_free()
126 if (likely(!xdp_buff_has_frags(xdp))) in xsk_buff_free()
134 xdp_get_shared_info_from_buff(xdp)->nr_frags = 0; in xsk_buff_free()
139 static inline void xsk_buff_add_frag(struct xdp_buff *xdp) in xsk_buff_add_frag() argument
141 struct xdp_buff_xsk *frag = container_of(xdp, struct xdp_buff_xsk, xdp); in xsk_buff_add_frag()
148 struct xdp_buff_xsk *xskb = container_of(first, struct xdp_buff_xsk, xdp); in xsk_buff_get_frag()
156 ret = &frag->xdp; in xsk_buff_get_frag()
164 struct xdp_buff_xsk *xskb = container_of(tail, struct xdp_buff_xsk, xdp); in xsk_buff_del_tail()
171 struct xdp_buff_xsk *xskb = container_of(first, struct xdp_buff_xsk, xdp); in xsk_buff_get_tail()
176 return &frag->xdp; in xsk_buff_get_tail()
179 static inline void xsk_buff_set_size(struct xdp_buff *xdp, u32 size) in xsk_buff_set_size() argument
181 xdp->data = xdp->data_hard_start + XDP_PACKET_HEADROOM; in xsk_buff_set_size()
182 xdp->data_meta = xdp->data; in xsk_buff_set_size()
183 xdp->data_end = xdp->data + size; in xsk_buff_set_size()
184 xdp->flags = 0; in xsk_buff_set_size()
222 static inline void xsk_buff_dma_sync_for_cpu(struct xdp_buff *xdp) in xsk_buff_dma_sync_for_cpu() argument
224 struct xdp_buff_xsk *xskb = container_of(xdp, struct xdp_buff_xsk, xdp); in xsk_buff_dma_sync_for_cpu()
325 static inline dma_addr_t xsk_buff_xdp_get_dma(struct xdp_buff *xdp) in xsk_buff_xdp_get_dma() argument
330 static inline dma_addr_t xsk_buff_xdp_get_frame_dma(struct xdp_buff *xdp) in xsk_buff_xdp_get_frame_dma() argument
345 static inline u32 xsk_buff_alloc_batch(struct xsk_buff_pool *pool, struct xdp_buff **xdp, u32 max) in xsk_buff_alloc_batch() argument
355 static inline void xsk_buff_free(struct xdp_buff *xdp) in xsk_buff_free() argument
359 static inline void xsk_buff_add_frag(struct xdp_buff *xdp) in xsk_buff_add_frag() argument
377 static inline void xsk_buff_set_size(struct xdp_buff *xdp, u32 size) in xsk_buff_set_size() argument
402 static inline void xsk_buff_dma_sync_for_cpu(struct xdp_buff *xdp) in xsk_buff_dma_sync_for_cpu() argument