12/10/2020 18:38, Andrew Rybchenko:
> On 10/12/20 7:19 PM, Viacheslav Ovsiienko wrote:
> >  int
> > +rte_eth_rxseg_queue_setup(uint16_t port_id, uint16_t rx_queue_id,
> > +                     uint16_t nb_rx_desc, unsigned int socket_id,
> > +                     const struct rte_eth_rxconf *rx_conf,
> > +                     const struct rte_eth_rxseg *rx_seg, uint16_t n_seg)
> > +{
> > +   int ret;
> > +   uint16_t seg_idx;
> > +   uint32_t mbp_buf_size;
> 
> <start-of-dup>
> 
> > +   struct rte_eth_dev *dev;
> > +   struct rte_eth_dev_info dev_info;
> > +   struct rte_eth_rxconf local_conf;
> > +   void **rxq;
> > +
> > +   RTE_ETH_VALID_PORTID_OR_ERR_RET(port_id, -EINVAL);
> > +
> > +   dev = &rte_eth_devices[port_id];
> > +   if (rx_queue_id >= dev->data->nb_rx_queues) {
> > +           RTE_ETHDEV_LOG(ERR, "Invalid RX queue_id=%u\n", rx_queue_id);
> > +           return -EINVAL;
> > +   }
> 
> <end-of-dup>
> 
> > +
> > +   if (rx_seg == NULL) {
> > +           RTE_ETHDEV_LOG(ERR, "Invalid null description pointer\n");
> > +           return -EINVAL;
> > +   }
> > +
> > +   if (n_seg == 0) {
> > +           RTE_ETHDEV_LOG(ERR, "Invalid zero description number\n");
> > +           return -EINVAL;
> > +   }
> > +
> > +   RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->rxseg_queue_setup, -ENOTSUP);
> > +
> 
> <start-of-dup>
> 
> > +   /*
> > +    * Check the size of the mbuf data buffer.
> > +    * This value must be provided in the private data of the memory pool.
> > +    * First check that the memory pool has a valid private data.
> > +    */
> > +   ret = rte_eth_dev_info_get(port_id, &dev_info);
> > +   if (ret != 0)
> > +           return ret;
> 
> <end-of-dup>
> 
> > +
> > +   for (seg_idx = 0; seg_idx < n_seg; seg_idx++) {
> > +           struct rte_mempool *mp = rx_seg[seg_idx].mp;
> > +
> > +           if (mp->private_data_size <
> > +                           sizeof(struct rte_pktmbuf_pool_private)) {
> > +                   RTE_ETHDEV_LOG(ERR, "%s private_data_size %d < %d\n",
> > +                           mp->name, (int)mp->private_data_size,
> > +                           (int)sizeof(struct rte_pktmbuf_pool_private));
> > +                   return -ENOSPC;
> > +           }
> > +
> > +           mbp_buf_size = rte_pktmbuf_data_room_size(mp);
> > +           if (mbp_buf_size < rx_seg[seg_idx].length +
> > +                              rx_seg[seg_idx].offset +
> > +                              (seg_idx ? 0 :
> > +                               (uint32_t)RTE_PKTMBUF_HEADROOM)) {
> > +                   RTE_ETHDEV_LOG(ERR,
> > +                           "%s mbuf_data_room_size %d < %d"
> > +                           " (segment length=%d + segment offset=%d)\n",
> > +                           mp->name, (int)mbp_buf_size,
> > +                           (int)(rx_seg[seg_idx].length +
> > +                                 rx_seg[seg_idx].offset),
> > +                           (int)rx_seg[seg_idx].length,
> > +                           (int)rx_seg[seg_idx].offset);
> > +                   return -EINVAL;
> > +           }
> > +   }
> > +
> 
> <start-of-huge-dup>
> 
> > +   /* Use default specified by driver, if nb_rx_desc is zero */
> > +   if (nb_rx_desc == 0) {
> > +           nb_rx_desc = dev_info.default_rxportconf.ring_size;
> > +           /* If driver default is also zero, fall back on EAL default */
> > +           if (nb_rx_desc == 0)
> > +                   nb_rx_desc = RTE_ETH_DEV_FALLBACK_RX_RINGSIZE;
> > +   }
> > +
> > +   if (nb_rx_desc > dev_info.rx_desc_lim.nb_max ||
> > +                   nb_rx_desc < dev_info.rx_desc_lim.nb_min ||
> > +                   nb_rx_desc % dev_info.rx_desc_lim.nb_align != 0) {
> > +
> > +           RTE_ETHDEV_LOG(ERR,
> > +                   "Invalid value for nb_rx_desc(=%hu), should be: "
> > +                   "<= %hu, >= %hu, and a product of %hu\n",
> > +                   nb_rx_desc, dev_info.rx_desc_lim.nb_max,
> > +                   dev_info.rx_desc_lim.nb_min,
> > +                   dev_info.rx_desc_lim.nb_align);
> > +           return -EINVAL;
> > +   }
> > +
> > +   if (dev->data->dev_started &&
> > +           !(dev_info.dev_capa &
> > +                   RTE_ETH_DEV_CAPA_RUNTIME_RX_QUEUE_SETUP))
> > +           return -EBUSY;
> > +
> > +   if (dev->data->dev_started &&
> > +           (dev->data->rx_queue_state[rx_queue_id] !=
> > +                   RTE_ETH_QUEUE_STATE_STOPPED))
> > +           return -EBUSY;
> > +
> > +   rxq = dev->data->rx_queues;
> > +   if (rxq[rx_queue_id]) {
> > +           RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->rx_queue_release,
> > +                                   -ENOTSUP);
> > +           (*dev->dev_ops->rx_queue_release)(rxq[rx_queue_id]);
> > +           rxq[rx_queue_id] = NULL;
> > +   }
> > +
> > +   if (rx_conf == NULL)
> > +           rx_conf = &dev_info.default_rxconf;
> > +
> > +   local_conf = *rx_conf;
> > +
> > +   /*
> > +    * If an offloading has already been enabled in
> > +    * rte_eth_dev_configure(), it has been enabled on all queues,
> > +    * so there is no need to enable it in this queue again.
> > +    * The local_conf.offloads input to underlying PMD only carries
> > +    * those offloadings which are only enabled on this queue and
> > +    * not enabled on all queues.
> > +    */
> > +   local_conf.offloads &= ~dev->data->dev_conf.rxmode.offloads;
> > +
> > +   /*
> > +    * New added offloadings for this queue are those not enabled in
> > +    * rte_eth_dev_configure() and they must be per-queue type.
> > +    * A pure per-port offloading can't be enabled on a queue while
> > +    * disabled on another queue. A pure per-port offloading can't
> > +    * be enabled for any queue as new added one if it hasn't been
> > +    * enabled in rte_eth_dev_configure().
> > +    */
> > +   if ((local_conf.offloads & dev_info.rx_queue_offload_capa) !=
> > +        local_conf.offloads) {
> > +           RTE_ETHDEV_LOG(ERR,
> > +                   "Ethdev port_id=%d rx_queue_id=%d, new added offloads"
> > +                   " 0x%"PRIx64" must be within per-queue offload"
> > +                   " capabilities 0x%"PRIx64" in %s()\n",
> > +                   port_id, rx_queue_id, local_conf.offloads,
> > +                   dev_info.rx_queue_offload_capa,
> > +                   __func__);
> > +           return -EINVAL;
> > +   }
> > +
> > +   /*
> > +    * If LRO is enabled, check that the maximum aggregated packet
> > +    * size is supported by the configured device.
> > +    */
> > +   if (local_conf.offloads & DEV_RX_OFFLOAD_TCP_LRO) {
> > +           if (dev->data->dev_conf.rxmode.max_lro_pkt_size == 0)
> > +                   dev->data->dev_conf.rxmode.max_lro_pkt_size =
> > +                           dev->data->dev_conf.rxmode.max_rx_pkt_len;
> > +           int ret = check_lro_pkt_size(port_id,
> > +                           dev->data->dev_conf.rxmode.max_lro_pkt_size,
> > +                           dev->data->dev_conf.rxmode.max_rx_pkt_len,
> > +                           dev_info.max_lro_pkt_size);
> > +           if (ret != 0)
> > +                   return ret;
> > +   }
> 
> <end-of-huge-dup>
> 
> IMO It is not acceptable to duplication so much code.
> It is simply unmaintainable.
> 
> NACK

Can it be solved by making rte_eth_rx_queue_setup() a wrapper
on top of this new rte_eth_rxseg_queue_setup() ?



Reply via email to