On Thu, Feb 21, 2019 at 12:22:29PM +0100, Maxime Coquelin wrote:
> On 2/19/19 11:59 AM, Tiwei Bie wrote:
> > This patch introduces an optimized enqueue function in packed
> > ring for the case that virtio net header can be prepended to
> > the unchained mbuf.
> > 
> > Signed-off-by: Tiwei Bie <tiwei....@intel.com>
> > ---
> >   drivers/net/virtio/virtio_rxtx.c | 63 +++++++++++++++++++++++++++++++-
> >   1 file changed, 61 insertions(+), 2 deletions(-)
> > 
> > diff --git a/drivers/net/virtio/virtio_rxtx.c 
> > b/drivers/net/virtio/virtio_rxtx.c
> > index 60fa3aa50..771d3c3f6 100644
> > --- a/drivers/net/virtio/virtio_rxtx.c
> > +++ b/drivers/net/virtio/virtio_rxtx.c
> > @@ -623,6 +623,62 @@ virtqueue_enqueue_xmit_inorder(struct virtnet_tx *txvq,
> >     vq->vq_desc_head_idx = idx & (vq->vq_nentries - 1);
> >   }
> > +static inline void
> > +virtqueue_enqueue_xmit_packed_fast(struct virtnet_tx *txvq,
> > +                              struct rte_mbuf *cookie,
> > +                              int in_order)
> > +{
> > +   struct virtqueue *vq = txvq->vq;
> > +   struct vring_packed_desc *dp;
> > +   struct vq_desc_extra *dxp;
> > +   uint16_t idx, id, flags;
> > +   uint16_t head_size = vq->hw->vtnet_hdr_size;
> > +   struct virtio_net_hdr *hdr;
> > +
> > +   id = in_order ? vq->vq_avail_idx : vq->vq_desc_head_idx;
> > +   idx = vq->vq_avail_idx;
> > +   dp = &vq->ring_packed.desc_packed[idx];
> > +
> > +   dxp = &vq->vq_descx[id];
> > +   dxp->ndescs = 1;
> > +   dxp->cookie = cookie;
> > +
> > +   flags = vq->avail_used_flags;
> > +
> > +   /* prepend cannot fail, checked by caller */
> > +   hdr = (struct virtio_net_hdr *)
> > +           rte_pktmbuf_prepend(cookie, head_size);
> > +   cookie->pkt_len -= head_size;
> > +
> > +   /* if offload disabled, hdr is not zeroed yet, do it now */
> > +   if (!vq->hw->has_tx_offload)
> > +           virtqueue_clear_net_hdr(hdr);
> > +   else
> > +           virtqueue_xmit_offload(hdr, cookie, true);
> > +
> > +   dp->addr = VIRTIO_MBUF_DATA_DMA_ADDR(cookie, vq);
> > +   dp->len  = cookie->data_len;
> > +   dp->id   = id;
> > +
> > +   if (++vq->vq_avail_idx >= vq->vq_nentries) {
> > +           vq->vq_avail_idx -= vq->vq_nentries;
> > +           vq->avail_wrap_counter ^= 1;
> > +           vq->avail_used_flags ^=
> > +                   VRING_DESC_F_AVAIL(1) | VRING_DESC_F_USED(1);
> > +   }
> > +
> > +   vq->vq_free_cnt--;
> > +
> > +   if (!in_order) {
> > +           vq->vq_desc_head_idx = dxp->next;
> > +           if (vq->vq_desc_head_idx == VQ_RING_DESC_CHAIN_END)
> > +                   vq->vq_desc_tail_idx = VQ_RING_DESC_CHAIN_END;
> > +   }
> > +
> > +   virtio_wmb(vq->hw->weak_barriers);
> > +   dp->flags = flags;
> > +}
> > +
> >   static inline void
> >   virtqueue_enqueue_xmit_packed(struct virtnet_tx *txvq, struct rte_mbuf 
> > *cookie,
> >                           uint16_t needed, int can_push, int in_order)
> > @@ -1979,8 +2035,11 @@ virtio_xmit_pkts_packed(void *tx_queue, struct 
> > rte_mbuf **tx_pkts,
> >             }
> >             /* Enqueue Packet buffers */
> > -           virtqueue_enqueue_xmit_packed(txvq, txm, slots, can_push,
> > -                                         in_order);
> > +           if (can_push)
> > +                   virtqueue_enqueue_xmit_packed_fast(txvq, txm, in_order);
> > +           else
> > +                   virtqueue_enqueue_xmit_packed(txvq, txm, slots, 0,
> > +                                                 in_order);
> >             virtio_update_packet_stats(&txvq->stats, txm);
> >     }
> > 
> 
> I like this patch, but shouldn't virtqueue_enqueue_xmit_packed() be
> simplified to get rid off "can_push" now that this case as a dedicated
> function?

Yeah, I had the same thought. But after a second thought, I
think we may also want to push the net hdr to the mbuf even
if its nb_segs isn't 1 in the future, so I left it untouched.

Thanks,
Tiwei

Reply via email to