Hi Ilya,

> Ping.
> 
> Best regards, Ilya Maximets.
> 
> On 19.10.2016 17:07, Ilya Maximets wrote:
> > The only reason why bulk alloc disabled for the rings with
> > more than (I40E_MAX_RING_DESC - RTE_PMD_I40E_RX_MAX_BURST)
> > descriptors is the possible out-of-bound access to the dma
> > memory. But it's the artificial limit and can be easily
> > avoided by allocating of RTE_PMD_I40E_RX_MAX_BURST more
> > descriptors in memory. This will not interfere the HW and,
> > as soon as all rings' memory zeroized, Rx functions will
> > work correctly.
> >
> > This change allows to use vectorized Rx functions with
> > 4096 descriptors in Rx ring which is important to achieve
> > zero packet drop rate in high-load installations.
> >
> > Signed-off-by: Ilya Maximets <i.maximets at samsung.com>
> > ---
> >  drivers/net/i40e/i40e_rxtx.c | 24 +++++++++++++-----------
> >  1 file changed, 13 insertions(+), 11 deletions(-)
> >
> > diff --git a/drivers/net/i40e/i40e_rxtx.c b/drivers/net/i40e/i40e_rxtx.c
> > index 7ae7d9f..1f76691 100644
> > --- a/drivers/net/i40e/i40e_rxtx.c
> > +++ b/drivers/net/i40e/i40e_rxtx.c
> > @@ -409,15 +409,6 @@ check_rx_burst_bulk_alloc_preconditions(__rte_unused 
> > struct i40e_rx_queue *rxq)
> >                          "rxq->rx_free_thresh=%d",
> >                          rxq->nb_rx_desc, rxq->rx_free_thresh);
> >             ret = -EINVAL;
> > -   } else if (!(rxq->nb_rx_desc < (I40E_MAX_RING_DESC -
> > -                           RTE_PMD_I40E_RX_MAX_BURST))) {
> > -           PMD_INIT_LOG(DEBUG, "Rx Burst Bulk Alloc Preconditions: "
> > -                        "rxq->nb_rx_desc=%d, "
> > -                        "I40E_MAX_RING_DESC=%d, "
> > -                        "RTE_PMD_I40E_RX_MAX_BURST=%d",
> > -                        rxq->nb_rx_desc, I40E_MAX_RING_DESC,
> > -                        RTE_PMD_I40E_RX_MAX_BURST);
> > -           ret = -EINVAL;
> >     }
> >  #else
> >     ret = -EINVAL;
> > @@ -1698,8 +1689,19 @@ i40e_dev_rx_queue_setup(struct rte_eth_dev *dev,
> >     rxq->rx_deferred_start = rx_conf->rx_deferred_start;
> >
> >     /* Allocate the maximun number of RX ring hardware descriptor. */
> > -   ring_size = sizeof(union i40e_rx_desc) * I40E_MAX_RING_DESC;
> > -   ring_size = RTE_ALIGN(ring_size, I40E_DMA_MEM_ALIGN);
> > +   len = I40E_MAX_RING_DESC;
> > +
> > +#ifdef RTE_LIBRTE_I40E_RX_ALLOW_BULK_ALLOC
> > +   /**
> > +    * Allocating a little more memory because vectorized/bulk_alloc Rx
> > +    * functions doesn't check boundaries each time.
> > +    */
> > +   len += RTE_PMD_I40E_RX_MAX_BURST;
> > +#endif
> > +

Looks good to me.
One question, though do we really need '+#ifdef 
RTE_LIBRTE_I40E_RX_ALLOW_BULK_ALLOC' here?
Why just not remove this ifdef here and always add allocate extra descriptors.
Konstantin

> > +   ring_size = RTE_ALIGN(len * sizeof(union i40e_rx_desc),
> > +                         I40E_DMA_MEM_ALIGN);
> > +
> >     rz = rte_eth_dma_zone_reserve(dev, "rx_ring", queue_idx,
> >                           ring_size, I40E_RING_BASE_ALIGN, socket_id);
> >     if (!rz) {
> >

Reply via email to