Complete the deduplication/merger of the ice and i40e Tx simple scalar
paths.

Signed-off-by: Bruce Richardson <[email protected]>
---
 drivers/net/intel/common/tx_scalar_fns.h | 87 ++++++++++++++++++++++++
 drivers/net/intel/i40e/i40e_rxtx.c       | 74 +-------------------
 drivers/net/intel/ice/ice_rxtx.c         | 74 +-------------------
 3 files changed, 89 insertions(+), 146 deletions(-)

diff --git a/drivers/net/intel/common/tx_scalar_fns.h 
b/drivers/net/intel/common/tx_scalar_fns.h
index d472aa24e0..dcdec200ac 100644
--- a/drivers/net/intel/common/tx_scalar_fns.h
+++ b/drivers/net/intel/common/tx_scalar_fns.h
@@ -127,6 +127,93 @@ ci_tx_free_bufs(struct ci_tx_queue *txq)
        return txq->tx_rs_thresh;
 }
 
+/* Simple burst transmit for descriptor-based simple Tx path
+ *
+ * Transmits a burst of packets by filling hardware descriptors with mbuf
+ * data. Handles ring wrap-around and RS bit management. Performs descriptor
+ * cleanup when tx_free_thresh is reached.
+ *
+ * Returns: number of packets transmitted
+ */
+static inline uint16_t
+ci_xmit_burst_simple(struct ci_tx_queue *txq,
+                    struct rte_mbuf **tx_pkts,
+                    uint16_t nb_pkts)
+{
+       volatile struct ci_tx_desc *txr = txq->ci_tx_ring;
+       uint16_t n = 0;
+
+       /**
+        * Begin scanning the H/W ring for done descriptors when the number
+        * of available descriptors drops below tx_free_thresh. For each done
+        * descriptor, free the associated buffer.
+        */
+       if (txq->nb_tx_free < txq->tx_free_thresh)
+               ci_tx_free_bufs(txq);
+
+       /* Use available descriptor only */
+       nb_pkts = (uint16_t)RTE_MIN(txq->nb_tx_free, nb_pkts);
+       if (unlikely(!nb_pkts))
+               return 0;
+
+       txq->nb_tx_free = (uint16_t)(txq->nb_tx_free - nb_pkts);
+       if ((txq->tx_tail + nb_pkts) > txq->nb_tx_desc) {
+               n = (uint16_t)(txq->nb_tx_desc - txq->tx_tail);
+               ci_tx_fill_hw_ring(txq, tx_pkts, n);
+               txr[txq->tx_next_rs].cmd_type_offset_bsz |=
+                       rte_cpu_to_le_64(((uint64_t)CI_TX_DESC_CMD_RS) <<
+                                         CI_TXD_QW1_CMD_S);
+               txq->tx_next_rs = (uint16_t)(txq->tx_rs_thresh - 1);
+               txq->tx_tail = 0;
+       }
+
+       /* Fill hardware descriptor ring with mbuf data */
+       ci_tx_fill_hw_ring(txq, tx_pkts + n, (uint16_t)(nb_pkts - n));
+       txq->tx_tail = (uint16_t)(txq->tx_tail + (nb_pkts - n));
+
+       /* Determine if RS bit needs to be set */
+       if (txq->tx_tail > txq->tx_next_rs) {
+               txr[txq->tx_next_rs].cmd_type_offset_bsz |=
+                       rte_cpu_to_le_64(((uint64_t)CI_TX_DESC_CMD_RS) <<
+                                         CI_TXD_QW1_CMD_S);
+               txq->tx_next_rs =
+                       (uint16_t)(txq->tx_next_rs + txq->tx_rs_thresh);
+               if (txq->tx_next_rs >= txq->nb_tx_desc)
+                       txq->tx_next_rs = (uint16_t)(txq->tx_rs_thresh - 1);
+       }
+
+       if (txq->tx_tail >= txq->nb_tx_desc)
+               txq->tx_tail = 0;
+
+       /* Update the tx tail register */
+       rte_write32_wc((uint32_t)txq->tx_tail, txq->qtx_tail);
+
+       return nb_pkts;
+}
+
+static __rte_always_inline uint16_t
+ci_xmit_pkts_simple(struct ci_tx_queue *txq,
+                    struct rte_mbuf **tx_pkts,
+                    uint16_t nb_pkts)
+{
+       uint16_t nb_tx = 0;
+
+       if (likely(nb_pkts <= CI_TX_MAX_BURST))
+               return ci_xmit_burst_simple(txq, tx_pkts, nb_pkts);
+
+       while (nb_pkts) {
+               uint16_t ret, num = RTE_MIN(nb_pkts, CI_TX_MAX_BURST);
+
+               ret = ci_xmit_burst_simple(txq, &tx_pkts[nb_tx], num);
+               nb_tx += ret;
+               nb_pkts -= ret;
+               if (ret < num)
+                       break;
+       }
+
+       return nb_tx;
+}
+
 /*
  * Common transmit descriptor cleanup function for Intel drivers.
  * Used by ice, i40e, iavf, and idpf drivers.
diff --git a/drivers/net/intel/i40e/i40e_rxtx.c 
b/drivers/net/intel/i40e/i40e_rxtx.c
index 22728af980..ac53554234 100644
--- a/drivers/net/intel/i40e/i40e_rxtx.c
+++ b/drivers/net/intel/i40e/i40e_rxtx.c
@@ -1012,84 +1012,12 @@ i40e_xmit_pkts(void *tx_queue, struct rte_mbuf 
**tx_pkts, uint16_t nb_pkts)
                        get_context_desc, NULL, NULL);
 }
 
-static inline uint16_t
-tx_xmit_pkts(struct ci_tx_queue *txq,
-            struct rte_mbuf **tx_pkts,
-            uint16_t nb_pkts)
-{
-       volatile struct ci_tx_desc *txr = txq->ci_tx_ring;
-       uint16_t n = 0;
-
-       /**
-        * Begin scanning the H/W ring for done descriptors when the number
-        * of available descriptors drops below tx_free_thresh. For each done
-        * descriptor, free the associated buffer.
-        */
-       if (txq->nb_tx_free < txq->tx_free_thresh)
-               ci_tx_free_bufs(txq);
-
-       /* Use available descriptor only */
-       nb_pkts = (uint16_t)RTE_MIN(txq->nb_tx_free, nb_pkts);
-       if (unlikely(!nb_pkts))
-               return 0;
-
-       txq->nb_tx_free = (uint16_t)(txq->nb_tx_free - nb_pkts);
-       if ((txq->tx_tail + nb_pkts) > txq->nb_tx_desc) {
-               n = (uint16_t)(txq->nb_tx_desc - txq->tx_tail);
-               ci_tx_fill_hw_ring(txq, tx_pkts, n);
-               txr[txq->tx_next_rs].cmd_type_offset_bsz |=
-                       rte_cpu_to_le_64(((uint64_t)CI_TX_DESC_CMD_RS) << 
CI_TXD_QW1_CMD_S);
-               txq->tx_next_rs = (uint16_t)(txq->tx_rs_thresh - 1);
-               txq->tx_tail = 0;
-       }
-
-       /* Fill hardware descriptor ring with mbuf data */
-       ci_tx_fill_hw_ring(txq, tx_pkts + n, (uint16_t)(nb_pkts - n));
-       txq->tx_tail = (uint16_t)(txq->tx_tail + (nb_pkts - n));
-
-       /* Determine if RS bit needs to be set */
-       if (txq->tx_tail > txq->tx_next_rs) {
-               txr[txq->tx_next_rs].cmd_type_offset_bsz |=
-                       rte_cpu_to_le_64(((uint64_t)CI_TX_DESC_CMD_RS) << 
CI_TXD_QW1_CMD_S);
-               txq->tx_next_rs =
-                       (uint16_t)(txq->tx_next_rs + txq->tx_rs_thresh);
-               if (txq->tx_next_rs >= txq->nb_tx_desc)
-                       txq->tx_next_rs = (uint16_t)(txq->tx_rs_thresh - 1);
-       }
-
-       if (txq->tx_tail >= txq->nb_tx_desc)
-               txq->tx_tail = 0;
-
-       /* Update the tx tail register */
-       I40E_PCI_REG_WC_WRITE(txq->qtx_tail, txq->tx_tail);
-
-       return nb_pkts;
-}
-
 static uint16_t
 i40e_xmit_pkts_simple(void *tx_queue,
                      struct rte_mbuf **tx_pkts,
                      uint16_t nb_pkts)
 {
-       uint16_t nb_tx = 0;
-
-       if (likely(nb_pkts <= CI_TX_MAX_BURST))
-               return tx_xmit_pkts((struct ci_tx_queue *)tx_queue,
-                                               tx_pkts, nb_pkts);
-
-       while (nb_pkts) {
-               uint16_t ret, num = (uint16_t)RTE_MIN(nb_pkts,
-                                               CI_TX_MAX_BURST);
-
-               ret = tx_xmit_pkts((struct ci_tx_queue *)tx_queue,
-                                               &tx_pkts[nb_tx], num);
-               nb_tx = (uint16_t)(nb_tx + ret);
-               nb_pkts = (uint16_t)(nb_pkts - ret);
-               if (ret < num)
-                       break;
-       }
-
-       return nb_tx;
+       return ci_xmit_pkts_simple(tx_queue, tx_pkts, nb_pkts);
 }
 
 #ifndef RTE_ARCH_X86
diff --git a/drivers/net/intel/ice/ice_rxtx.c b/drivers/net/intel/ice/ice_rxtx.c
index 468c039ab1..ed82a84dc5 100644
--- a/drivers/net/intel/ice/ice_rxtx.c
+++ b/drivers/net/intel/ice/ice_rxtx.c
@@ -3235,84 +3235,12 @@ ice_tx_done_cleanup(void *txq, uint32_t free_cnt)
                return ice_tx_done_cleanup_full(q, free_cnt);
 }
 
-static inline uint16_t
-tx_xmit_pkts(struct ci_tx_queue *txq,
-            struct rte_mbuf **tx_pkts,
-            uint16_t nb_pkts)
-{
-       volatile struct ci_tx_desc *txr = txq->ci_tx_ring;
-       uint16_t n = 0;
-
-       /**
-        * Begin scanning the H/W ring for done descriptors when the number
-        * of available descriptors drops below tx_free_thresh. For each done
-        * descriptor, free the associated buffer.
-        */
-       if (txq->nb_tx_free < txq->tx_free_thresh)
-               ci_tx_free_bufs(txq);
-
-       /* Use available descriptor only */
-       nb_pkts = (uint16_t)RTE_MIN(txq->nb_tx_free, nb_pkts);
-       if (unlikely(!nb_pkts))
-               return 0;
-
-       txq->nb_tx_free = (uint16_t)(txq->nb_tx_free - nb_pkts);
-       if ((txq->tx_tail + nb_pkts) > txq->nb_tx_desc) {
-               n = (uint16_t)(txq->nb_tx_desc - txq->tx_tail);
-               ci_tx_fill_hw_ring(txq, tx_pkts, n);
-               txr[txq->tx_next_rs].cmd_type_offset_bsz |=
-                       rte_cpu_to_le_64(((uint64_t)CI_TX_DESC_CMD_RS) << 
CI_TXD_QW1_CMD_S);
-               txq->tx_next_rs = (uint16_t)(txq->tx_rs_thresh - 1);
-               txq->tx_tail = 0;
-       }
-
-       /* Fill hardware descriptor ring with mbuf data */
-       ci_tx_fill_hw_ring(txq, tx_pkts + n, (uint16_t)(nb_pkts - n));
-       txq->tx_tail = (uint16_t)(txq->tx_tail + (nb_pkts - n));
-
-       /* Determine if RS bit needs to be set */
-       if (txq->tx_tail > txq->tx_next_rs) {
-               txr[txq->tx_next_rs].cmd_type_offset_bsz |=
-                       rte_cpu_to_le_64(((uint64_t)CI_TX_DESC_CMD_RS) << 
CI_TXD_QW1_CMD_S);
-               txq->tx_next_rs =
-                       (uint16_t)(txq->tx_next_rs + txq->tx_rs_thresh);
-               if (txq->tx_next_rs >= txq->nb_tx_desc)
-                       txq->tx_next_rs = (uint16_t)(txq->tx_rs_thresh - 1);
-       }
-
-       if (txq->tx_tail >= txq->nb_tx_desc)
-               txq->tx_tail = 0;
-
-       /* Update the tx tail register */
-       ICE_PCI_REG_WC_WRITE(txq->qtx_tail, txq->tx_tail);
-
-       return nb_pkts;
-}
-
 static uint16_t
 ice_xmit_pkts_simple(void *tx_queue,
                     struct rte_mbuf **tx_pkts,
                     uint16_t nb_pkts)
 {
-       uint16_t nb_tx = 0;
-
-       if (likely(nb_pkts <= CI_TX_MAX_BURST))
-               return tx_xmit_pkts((struct ci_tx_queue *)tx_queue,
-                                   tx_pkts, nb_pkts);
-
-       while (nb_pkts) {
-               uint16_t ret, num = (uint16_t)RTE_MIN(nb_pkts,
-                                                     CI_TX_MAX_BURST);
-
-               ret = tx_xmit_pkts((struct ci_tx_queue *)tx_queue,
-                                  &tx_pkts[nb_tx], num);
-               nb_tx = (uint16_t)(nb_tx + ret);
-               nb_pkts = (uint16_t)(nb_pkts - ret);
-               if (ret < num)
-                       break;
-       }
-
-       return nb_tx;
+       return ci_xmit_pkts_simple(tx_queue, tx_pkts, nb_pkts);
 }
 
 static const struct ci_rx_path_info ice_rx_path_infos[] = {
-- 
2.51.0

Reply via email to