From: Zaremba, Larysa <[email protected]> Date: Mon, 16 Feb 2026 15:01:45 +0100
> On Mon, Feb 16, 2026 at 11:48:27AM +0100, Alexander Lobakin wrote: >> From: Alexander Lobakin <[email protected]> >> Date: Mon, 16 Feb 2026 11:46:05 +0100 >> >>> From: Zaremba, Larysa <[email protected]> >>> Date: Thu, 12 Feb 2026 19:33:22 +0100 >>> >>>> The only user of frag_size field in XDP RxQ info is >>>> bpf_xdp_frags_increase_tail(). It clearly expects whole buffer size instead >>>> of DMA write size. Different assumptions in idpf driver configuration lead >>>> to negative tailroom. >>>> >>>> To make it worse, buffer sizes are not actually uniform in idpf when >>>> splitq is enabled, as there are several buffer queues, so rxq->rx_buf_size >>>> is meaningless in this case. >>>> >>>> Use rxq->truesize as a frag_size for singleq and truesize of the first bufq >>>> in AF_XDP ZC, as there is only one. Disable growinf tail for regular >>>> splitq. >>>> >>>> Fixes: ac8a861f632e ("idpf: prepare structures to support XDP") >>>> Reviewed-by: Aleksandr Loktionov <[email protected]> >>>> Signed-off-by: Larysa Zaremba <[email protected]> >>>> --- >>>> drivers/net/ethernet/intel/idpf/xdp.c | 8 +++++++- >>>> drivers/net/ethernet/intel/idpf/xsk.c | 1 + >>>> 2 files changed, 8 insertions(+), 1 deletion(-) >>>> >>>> diff --git a/drivers/net/ethernet/intel/idpf/xdp.c >>>> b/drivers/net/ethernet/intel/idpf/xdp.c >>>> index 958d16f87424..a152c9a26976 100644 >>>> --- a/drivers/net/ethernet/intel/idpf/xdp.c >>>> +++ b/drivers/net/ethernet/intel/idpf/xdp.c >>>> @@ -46,11 +46,17 @@ static int __idpf_xdp_rxq_info_init(struct >>>> idpf_rx_queue *rxq, void *arg) >>>> { >>>> const struct idpf_vport *vport = rxq->q_vector->vport; >>>> bool split = idpf_is_queue_model_split(vport->rxq_model); >>>> + u32 frag_size = 0; >>>> int err; >>>> >>>> + if (idpf_queue_has(XSK, rxq) && split) >>>> + frag_size = rxq->bufq_sets[0].bufq.truesize; >>>> + else if (!split) >>>> + frag_size = rxq->truesize; >>> >>> XDP and XSk are supported only in mode splitq mode, so you can remove >>> the second condition and change the first one to just `has(XSK)`. >>> > > But the function itself handles singleq case. I do not see why frag_size > should > be treated differently. singleq will still be handled and will always receive 0 here as in case when XSk is not enabled. > > Not that I am against of removing this logic, it would look more neat without > these conditions. > >>>> + >>>> err = __xdp_rxq_info_reg(&rxq->xdp_rxq, vport->netdev, rxq->idx, >>>> rxq->q_vector->napi.napi_id, >>>> - rxq->rx_buf_size); >>>> + frag_size); >>>> if (err) >>>> return err; >>>> >>>> diff --git a/drivers/net/ethernet/intel/idpf/xsk.c >>>> b/drivers/net/ethernet/intel/idpf/xsk.c >>>> index fd2cc43ab43c..febe1073b9b4 100644 >>>> --- a/drivers/net/ethernet/intel/idpf/xsk.c >>>> +++ b/drivers/net/ethernet/intel/idpf/xsk.c >>>> @@ -401,6 +401,7 @@ int idpf_xskfq_init(struct idpf_buf_queue *bufq) >>>> bufq->pending = fq.pending; >>>> bufq->thresh = fq.thresh; >>>> bufq->rx_buf_size = fq.buf_len; >>>> + bufq->truesize = xsk_pool_get_rx_frag_step(fq.pool); >> >> Better to do that in libeth_xdp rather than here? >> > > Smth like that? > > diff --git a/drivers/net/ethernet/intel/idpf/xsk.c > b/drivers/net/ethernet/intel/idpf/xsk.c > index febe1073b9b4..de87455b92d7 100644 > --- a/drivers/net/ethernet/intel/idpf/xsk.c > +++ b/drivers/net/ethernet/intel/idpf/xsk.c > @@ -401,7 +401,7 @@ int idpf_xskfq_init(struct idpf_buf_queue *bufq) > bufq->pending = fq.pending; > bufq->thresh = fq.thresh; > bufq->rx_buf_size = fq.buf_len; > - bufq->truesize = xsk_pool_get_rx_frag_step(fq.pool); > + bufq->truesize = fq.chunk_align; > > if (!idpf_xskfq_refill(bufq)) > netdev_err(bufq->pool->netdev, > diff --git a/drivers/net/ethernet/intel/libeth/xsk.c > b/drivers/net/ethernet/intel/libeth/xsk.c > index 846e902e31b6..5b298558ecfd 100644 > --- a/drivers/net/ethernet/intel/libeth/xsk.c > +++ b/drivers/net/ethernet/intel/libeth/xsk.c > @@ -167,6 +167,7 @@ int libeth_xskfq_create(struct libeth_xskfq *fq) > fq->pending = fq->count; > fq->thresh = libeth_xdp_queue_threshold(fq->count); > fq->buf_len = xsk_pool_get_rx_frame_size(fq->pool); > + fq->chunk_align = xsk_pool_get_rx_frag_step(fq->pool); > > return 0; > } > diff --git a/include/net/libeth/xsk.h b/include/net/libeth/xsk.h > index 481a7b28e6f2..a3ea90d30d17 100644 > --- a/include/net/libeth/xsk.h > +++ b/include/net/libeth/xsk.h > @@ -598,6 +598,7 @@ __libeth_xsk_run_pass(struct libeth_xdp_buff *xdp, > * @thresh: threshold below which the queue is refilled > * @buf_len: HW-writeable length per each buffer > * @nid: ID of the closest NUMA node with memory > + * @chunk_align: step between consecutive buffers, 0 if none exists > */ > struct libeth_xskfq { > struct_group_tagged(libeth_xskfq_fp, fp, > @@ -615,6 +616,8 @@ struct libeth_xskfq { > > u32 buf_len; > int nid; > + > + u32 chunk_align; > }; > > int libeth_xskfq_create(struct libeth_xskfq *fq); Ah, I forgot that for XSkFQ we have a separate structure... I'd do it like that: u32 buf_len; u32 truesize; // new field int nid; }; Maybe it's not really truesize, but it will be more straightforward and clear to the users what it means and which field they should assign the value to. truesize/chunk_align is related closely to buf_len, so I'd group them together in the struct, while nid is a different story, so it might stay separated by a newline. > > >>>> >>>> if (!idpf_xskfq_refill(bufq)) >>>> netdev_err(bufq->pool->netdev, Thanks, Olek

