virtio_dev_rx & scatter_rx & merge-able rx should be merged and the code
could be much simpler, unless there is special performance consideration.


On 5/28/2015 11:17 PM, Ouyang, Changchun wrote:
> Add support copying scattered mbuf to vring which is done by dev_scatter_rx,
> and check the 'next' pointer in mbuf on the fly to select suitable function 
> to rx packets.
>
> Signed-off-by: Changchun Ouyang <changchun.ouyang at intel.com>
> ---
>  lib/librte_vhost/vhost_rxtx.c | 116 
> +++++++++++++++++++++++++++++++++++++++++-
>  1 file changed, 115 insertions(+), 1 deletion(-)
>
> diff --git a/lib/librte_vhost/vhost_rxtx.c b/lib/librte_vhost/vhost_rxtx.c
> index bb56ae1..3086bb4 100644
> --- a/lib/librte_vhost/vhost_rxtx.c
> +++ b/lib/librte_vhost/vhost_rxtx.c
> @@ -46,7 +46,8 @@
>   * This function adds buffers to the virtio devices RX virtqueue. Buffers can
>   * be received from the physical port or from another virtio device. A packet
>   * count is returned to indicate the number of packets that are succesfully
> - * added to the RX queue. This function works when mergeable is disabled.
> + * added to the RX queue. This function works when mergeable is disabled and
> + * the mbuf is not scattered.
>   */
>  static inline uint32_t __attribute__((always_inline))
>  virtio_dev_rx(struct virtio_net *dev, uint16_t queue_id,
> @@ -447,6 +448,103 @@ fill_buf_vec(struct vhost_virtqueue *vq, uint16_t id, 
> uint32_t *vec_idx)
>  }
>  
>  /*
> + * This function works for scatter-gather RX.
> + */
> +static inline uint32_t __attribute__((always_inline))
> +virtio_dev_scatter_rx(struct virtio_net *dev, uint16_t queue_id,
> +     struct rte_mbuf **pkts, uint32_t count)
> +{
> +     struct vhost_virtqueue *vq;
> +     uint32_t pkt_idx = 0, entry_success = 0;
> +     uint16_t avail_idx;
> +     uint16_t res_base_idx, res_end_idx;
> +     uint8_t success = 0;
> +
> +     LOG_DEBUG(VHOST_DATA, "(%"PRIu64") virtio_dev_scatter_rx()\n",
> +             dev->device_fh);
use __func__
> +     if (unlikely(queue_id != VIRTIO_RXQ))
> +             LOG_DEBUG(VHOST_DATA, "mq isn't supported in this version.\n");
> +
> +     vq = dev->virtqueue[VIRTIO_RXQ];
> +     count = RTE_MIN((uint32_t)MAX_PKT_BURST, count);
> +
> +     if (count == 0)
> +             return 0;
> +
> +     for (pkt_idx = 0; pkt_idx < count; pkt_idx++) {
> +             uint32_t secure_len = 0;
> +             uint32_t vec_idx = 0;
> +             uint32_t pkt_len = pkts[pkt_idx]->pkt_len + vq->vhost_hlen;
> +
> +             do {
> +                     /*
> +                      * As many data cores may want access to available
> +                      * buffers, they need to be reserved.
> +                      */
> +                     res_base_idx = vq->last_used_idx_res;
> +                     avail_idx = *((volatile uint16_t *)&vq->avail->idx);
> +
> +                     if (unlikely(res_base_idx == avail_idx)) {
> +                             LOG_DEBUG(VHOST_DATA,
> +                                     "(%"PRIu64") Failed "
> +                                     "to get enough desc from "
> +                                     "vring\n",
> +                                     dev->device_fh);
> +                             return pkt_idx;
> +                     } else {
> +                             uint16_t wrapped_idx =
> +                                     (res_base_idx) & (vq->size - 1);
> +                             uint32_t idx = vq->avail->ring[wrapped_idx];
> +
> +                             update_secure_len(vq, idx, &secure_len);
> +                     }
> +
> +                     if (pkt_len > secure_len) {
> +                             LOG_DEBUG(VHOST_DATA,
> +                                     "(%"PRIu64") Failed "
> +                                     "to get enough desc from "
> +                                     "vring\n",
> +                                     dev->device_fh);
> +                             return pkt_idx;
> +                     }
The behavior for virtio_dev_rx and virtio_dev_merge_rx is totally
different. I think they should behave in the same way.
virtio_dev_rx updates used->len to zero while this one returns immediately.

Besides, with this implementation, if the caller retransmit the
mbuf(which has pkt_len larger the secure_len), it will enter endless loop.

> +
> +                     /* vq->last_used_idx_res is atomically updated. */
> +                     success = rte_atomic16_cmpset(&vq->last_used_idx_res,
> +                                                     res_base_idx,
> +                                                     res_base_idx + 1);
> +             } while (success == 0);

Here the behavior becomes different again in reserving vring entries.

> +
> +             fill_buf_vec(vq, res_base_idx, &vec_idx);
> +
> +             res_end_idx = res_base_idx + 1;
> +
> +             entry_success = copy_from_mbuf_to_vring(dev, res_base_idx,
> +                     res_end_idx, pkts[pkt_idx]);
> +
> +             rte_compiler_barrier();
> +
> +             /*
> +              * Wait until it's our turn to add our buffer
> +              * to the used ring.
> +              */
> +             while (unlikely(vq->last_used_idx != res_base_idx))
> +                     rte_pause();
> +
> +             *(volatile uint16_t *)&vq->used->idx += entry_success;
> +             vq->last_used_idx = res_end_idx;
> +
> +             /* flush used->idx update before we read avail->flags. */
> +             rte_mb();
> +
> +             /* Kick the guest if necessary. */
> +             if (!(vq->avail->flags & VRING_AVAIL_F_NO_INTERRUPT))
> +                     eventfd_write((int)vq->callfd, 1);
> +     }
> +
> +     return count;
> +}
> +
> +/*
>   * This function works for mergeable RX.
>   */
>  static inline uint32_t __attribute__((always_inline))
> @@ -545,12 +643,28 @@ virtio_dev_merge_rx(struct virtio_net *dev, uint16_t 
> queue_id,
>       return count;
>  }
>  
> +/*
> + * Return 1 if any mbuf is scattered, otherwise return 0.
> + */
> +static inline uint32_t __attribute__((always_inline))
> +check_scatter(struct rte_mbuf **pkts, uint16_t count)
> +{
> +     uint32_t i;
> +     for (i = 0; i < count; i++) {
> +             if (pkts[i]->next != NULL)
> +                     return 1;
> +     }
> +     return 0;
> +}
> +
>  uint16_t
>  rte_vhost_enqueue_burst(struct virtio_net *dev, uint16_t queue_id,
>       struct rte_mbuf **pkts, uint16_t count)
>  {
>       if (unlikely(dev->features & (1 << VIRTIO_NET_F_MRG_RXBUF)))
>               return virtio_dev_merge_rx(dev, queue_id, pkts, count);
> +     else if (unlikely(check_scatter(pkts, count) == 1))
> +             return virtio_dev_scatter_rx(dev, queue_id, pkts, count);
>       else
>               return virtio_dev_rx(dev, queue_id, pkts, count);
>  }

Reply via email to