Multiple drivers used the same logic to calculate how many Tx data
descriptors were needed. Move that calculation to common code. In the
process of updating drivers, fix idpf driver calculation for the TSO
case.

Signed-off-by: Bruce Richardson <[email protected]>
---
 drivers/net/intel/common/tx_scalar_fns.h  | 21 +++++++++++++++++++++
 drivers/net/intel/i40e/i40e_rxtx.c        | 18 +-----------------
 drivers/net/intel/iavf/iavf_rxtx.c        | 17 +----------------
 drivers/net/intel/ice/ice_rxtx.c          | 18 +-----------------
 drivers/net/intel/idpf/idpf_common_rxtx.c | 21 +++++++++++++++++----
 5 files changed, 41 insertions(+), 54 deletions(-)

diff --git a/drivers/net/intel/common/tx_scalar_fns.h 
b/drivers/net/intel/common/tx_scalar_fns.h
index c79210d084..f894cea616 100644
--- a/drivers/net/intel/common/tx_scalar_fns.h
+++ b/drivers/net/intel/common/tx_scalar_fns.h
@@ -64,4 +64,25 @@ ci_tx_xmit_cleanup(struct ci_tx_queue *txq)
        return 0;
 }
 
+static inline uint16_t
+ci_div_roundup16(uint16_t x, uint16_t y)
+{
+       return (uint16_t)((x + y - 1) / y);
+}
+
+/* Calculate the number of TX descriptors needed for each pkt */
+static inline uint16_t
+ci_calc_pkt_desc(const struct rte_mbuf *tx_pkt)
+{
+       uint16_t count = 0;
+
+       while (tx_pkt != NULL) {
+               count += ci_div_roundup16(tx_pkt->data_len, 
CI_MAX_DATA_PER_TXD);
+               tx_pkt = tx_pkt->next;
+       }
+
+       return count;
+}
+
+
 #endif /* _COMMON_INTEL_TX_SCALAR_FNS_H_ */
diff --git a/drivers/net/intel/i40e/i40e_rxtx.c 
b/drivers/net/intel/i40e/i40e_rxtx.c
index 892069372f..886be06a89 100644
--- a/drivers/net/intel/i40e/i40e_rxtx.c
+++ b/drivers/net/intel/i40e/i40e_rxtx.c
@@ -1025,21 +1025,6 @@ i40e_set_tso_ctx(struct rte_mbuf *mbuf, union 
ci_tx_offload tx_offload)
        return ctx_desc;
 }
 
-/* Calculate the number of TX descriptors needed for each pkt */
-static inline uint16_t
-i40e_calc_pkt_desc(struct rte_mbuf *tx_pkt)
-{
-       struct rte_mbuf *txd = tx_pkt;
-       uint16_t count = 0;
-
-       while (txd != NULL) {
-               count += DIV_ROUND_UP(txd->data_len, CI_MAX_DATA_PER_TXD);
-               txd = txd->next;
-       }
-
-       return count;
-}
-
 uint16_t
 i40e_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, uint16_t nb_pkts)
 {
@@ -1102,8 +1087,7 @@ i40e_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, 
uint16_t nb_pkts)
                 * per tx desc.
                 */
                if (ol_flags & RTE_MBUF_F_TX_TCP_SEG)
-                       nb_used = (uint16_t)(i40e_calc_pkt_desc(tx_pkt) +
-                                            nb_ctx);
+                       nb_used = (uint16_t)(ci_calc_pkt_desc(tx_pkt) + nb_ctx);
                else
                        nb_used = (uint16_t)(tx_pkt->nb_segs + nb_ctx);
                tx_last = (uint16_t)(tx_id + nb_used - 1);
diff --git a/drivers/net/intel/iavf/iavf_rxtx.c 
b/drivers/net/intel/iavf/iavf_rxtx.c
index 9946e112e8..ecf954a2c2 100644
--- a/drivers/net/intel/iavf/iavf_rxtx.c
+++ b/drivers/net/intel/iavf/iavf_rxtx.c
@@ -2667,21 +2667,6 @@ iavf_build_data_desc_cmd_offset_fields(volatile uint64_t 
*qw1,
                ((uint64_t)l2tag1 << IAVF_TXD_DATA_QW1_L2TAG1_SHIFT));
 }
 
-/* Calculate the number of TX descriptors needed for each pkt */
-static inline uint16_t
-iavf_calc_pkt_desc(struct rte_mbuf *tx_pkt)
-{
-       struct rte_mbuf *txd = tx_pkt;
-       uint16_t count = 0;
-
-       while (txd != NULL) {
-               count += (txd->data_len + CI_MAX_DATA_PER_TXD - 1) / 
CI_MAX_DATA_PER_TXD;
-               txd = txd->next;
-       }
-
-       return count;
-}
-
 static inline void
 iavf_fill_data_desc(volatile struct ci_tx_desc *desc,
        uint64_t desc_template, uint16_t buffsz,
@@ -2767,7 +2752,7 @@ iavf_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, 
uint16_t nb_pkts)
                 * per tx desc.
                 */
                if (mb->ol_flags & RTE_MBUF_F_TX_TCP_SEG)
-                       nb_desc_required = iavf_calc_pkt_desc(mb) + nb_desc_ctx 
+ nb_desc_ipsec;
+                       nb_desc_required = ci_calc_pkt_desc(mb) + nb_desc_ctx + 
nb_desc_ipsec;
                else
                        nb_desc_required = nb_desc_data + nb_desc_ctx + 
nb_desc_ipsec;
 
diff --git a/drivers/net/intel/ice/ice_rxtx.c b/drivers/net/intel/ice/ice_rxtx.c
index 5864238092..c2a38b1a13 100644
--- a/drivers/net/intel/ice/ice_rxtx.c
+++ b/drivers/net/intel/ice/ice_rxtx.c
@@ -3067,21 +3067,6 @@ ice_set_tso_ctx(struct rte_mbuf *mbuf, union 
ci_tx_offload tx_offload)
        return ctx_desc;
 }
 
-/* Calculate the number of TX descriptors needed for each pkt */
-static inline uint16_t
-ice_calc_pkt_desc(struct rte_mbuf *tx_pkt)
-{
-       struct rte_mbuf *txd = tx_pkt;
-       uint16_t count = 0;
-
-       while (txd != NULL) {
-               count += DIV_ROUND_UP(txd->data_len, CI_MAX_DATA_PER_TXD);
-               txd = txd->next;
-       }
-
-       return count;
-}
-
 uint16_t
 ice_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, uint16_t nb_pkts)
 {
@@ -3144,8 +3129,7 @@ ice_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, 
uint16_t nb_pkts)
                 * per tx desc.
                 */
                if (ol_flags & (RTE_MBUF_F_TX_TCP_SEG | RTE_MBUF_F_TX_UDP_SEG))
-                       nb_used = (uint16_t)(ice_calc_pkt_desc(tx_pkt) +
-                                            nb_ctx);
+                       nb_used = (uint16_t)(ci_calc_pkt_desc(tx_pkt) + nb_ctx);
                else
                        nb_used = (uint16_t)(tx_pkt->nb_segs + nb_ctx);
                tx_last = (uint16_t)(tx_id + nb_used - 1);
diff --git a/drivers/net/intel/idpf/idpf_common_rxtx.c 
b/drivers/net/intel/idpf/idpf_common_rxtx.c
index 587871b54a..11d6848430 100644
--- a/drivers/net/intel/idpf/idpf_common_rxtx.c
+++ b/drivers/net/intel/idpf/idpf_common_rxtx.c
@@ -934,7 +934,16 @@ idpf_dp_splitq_xmit_pkts(void *tx_queue, struct rte_mbuf 
**tx_pkts,
                tx_offload.tso_segsz = tx_pkt->tso_segsz;
                /* Calculate the number of context descriptors needed. */
                nb_ctx = idpf_calc_context_desc(ol_flags);
-               nb_used = tx_pkt->nb_segs + nb_ctx;
+
+               /* Calculate the number of TX descriptors needed for
+                * each packet. For TSO packets, use ci_calc_pkt_desc as
+                * the mbuf data size might exceed max data size that hw allows
+                * per tx desc.
+                */
+               if (ol_flags & RTE_MBUF_F_TX_TCP_SEG)
+                       nb_used = ci_calc_pkt_desc(tx_pkt) + nb_ctx;
+               else
+                       nb_used = tx_pkt->nb_segs + nb_ctx;
 
                if (ol_flags & IDPF_TX_CKSUM_OFFLOAD_MASK)
                        cmd_dtype = IDPF_TXD_FLEX_FLOW_CMD_CS_EN;
@@ -1382,10 +1391,14 @@ idpf_dp_singleq_xmit_pkts(void *tx_queue, struct 
rte_mbuf **tx_pkts,
                nb_ctx = idpf_calc_context_desc(ol_flags);
 
                /* The number of descriptors that must be allocated for
-                * a packet equals to the number of the segments of that
-                * packet plus 1 context descriptor if needed.
+                * a packet. For TSO packets, use ci_calc_pkt_desc as
+                * the mbuf data size might exceed max data size that hw allows
+                * per tx desc.
                 */
-               nb_used = (uint16_t)(tx_pkt->nb_segs + nb_ctx);
+               if (ol_flags & RTE_MBUF_F_TX_TCP_SEG)
+                       nb_used = (uint16_t)(ci_calc_pkt_desc(tx_pkt) + nb_ctx);
+               else
+                       nb_used = (uint16_t)(tx_pkt->nb_segs + nb_ctx);
                tx_last = (uint16_t)(tx_id + nb_used - 1);
 
                /* Circular ring */
-- 
2.51.0

Reply via email to