Similar thing has been done in i40e, as there is no real need for having
the sk_buff pointer in each rx_buf. Non-eop frames can be simply handled
on that pointer moved upwards to rx_ring.

Signed-off-by: Maciej Fijalkowski <maciej.fijalkow...@intel.com>
---
 drivers/net/ethernet/intel/ice/ice_txrx.c | 29 ++++++++++-------------
 drivers/net/ethernet/intel/ice/ice_txrx.h |  2 +-
 2 files changed, 13 insertions(+), 18 deletions(-)

diff --git a/drivers/net/ethernet/intel/ice/ice_txrx.c 
b/drivers/net/ethernet/intel/ice/ice_txrx.c
index 8b5d23436904..0ad89ee48c09 100644
--- a/drivers/net/ethernet/intel/ice/ice_txrx.c
+++ b/drivers/net/ethernet/intel/ice/ice_txrx.c
@@ -375,6 +375,11 @@ void ice_clean_rx_ring(struct ice_ring *rx_ring)
        if (!rx_ring->rx_buf)
                return;
 
+       if (rx_ring->skb) {
+               dev_kfree_skb(rx_ring->skb);
+               rx_ring->skb = NULL;
+       }
+
        if (rx_ring->xsk_pool) {
                ice_xsk_clean_rx_ring(rx_ring);
                goto rx_skip_free;
@@ -384,10 +389,6 @@ void ice_clean_rx_ring(struct ice_ring *rx_ring)
        for (i = 0; i < rx_ring->count; i++) {
                struct ice_rx_buf *rx_buf = &rx_ring->rx_buf[i];
 
-               if (rx_buf->skb) {
-                       dev_kfree_skb(rx_buf->skb);
-                       rx_buf->skb = NULL;
-               }
                if (!rx_buf->page)
                        continue;
 
@@ -857,21 +858,18 @@ ice_reuse_rx_page(struct ice_ring *rx_ring, struct 
ice_rx_buf *old_buf)
 /**
  * ice_get_rx_buf - Fetch Rx buffer and synchronize data for use
  * @rx_ring: Rx descriptor ring to transact packets on
- * @skb: skb to be used
  * @size: size of buffer to add to skb
  *
  * This function will pull an Rx buffer from the ring and synchronize it
  * for use by the CPU.
  */
 static struct ice_rx_buf *
-ice_get_rx_buf(struct ice_ring *rx_ring, struct sk_buff **skb,
-              const unsigned int size)
+ice_get_rx_buf(struct ice_ring *rx_ring, const unsigned int size)
 {
        struct ice_rx_buf *rx_buf;
 
        rx_buf = &rx_ring->rx_buf[rx_ring->next_to_clean];
        prefetchw(rx_buf->page);
-       *skb = rx_buf->skb;
 
        if (!size)
                return rx_buf;
@@ -1030,29 +1028,24 @@ static void ice_put_rx_buf(struct ice_ring *rx_ring, 
struct ice_rx_buf *rx_buf)
 
        /* clear contents of buffer_info */
        rx_buf->page = NULL;
-       rx_buf->skb = NULL;
 }
 
 /**
  * ice_is_non_eop - process handling of non-EOP buffers
  * @rx_ring: Rx ring being processed
  * @rx_desc: Rx descriptor for current buffer
- * @skb: Current socket buffer containing buffer in progress
  *
  * If the buffer is an EOP buffer, this function exits returning false,
  * otherwise return true indicating that this is in fact a non-EOP buffer.
  */
 static bool
-ice_is_non_eop(struct ice_ring *rx_ring, union ice_32b_rx_flex_desc *rx_desc,
-              struct sk_buff *skb)
+ice_is_non_eop(struct ice_ring *rx_ring, union ice_32b_rx_flex_desc *rx_desc)
 {
        /* if we are the last buffer then there is nothing else to do */
 #define ICE_RXD_EOF BIT(ICE_RX_FLEX_DESC_STATUS0_EOF_S)
        if (likely(ice_test_staterr(rx_desc, ICE_RXD_EOF)))
                return false;
 
-       /* place skb in next buffer to be received */
-       rx_ring->rx_buf[rx_ring->next_to_clean].skb = skb;
        rx_ring->rx_stats.non_eop_descs++;
 
        return true;
@@ -1075,6 +1068,7 @@ int ice_clean_rx_irq(struct ice_ring *rx_ring, int budget)
        unsigned int total_rx_bytes = 0, total_rx_pkts = 0;
        u16 cleaned_count = ICE_DESC_UNUSED(rx_ring);
        unsigned int xdp_res, xdp_xmit = 0;
+       struct sk_buff *skb = rx_ring->skb;
        struct bpf_prog *xdp_prog = NULL;
        struct xdp_buff xdp;
        bool failure;
@@ -1089,7 +1083,6 @@ int ice_clean_rx_irq(struct ice_ring *rx_ring, int budget)
        while (likely(total_rx_pkts < (unsigned int)budget)) {
                union ice_32b_rx_flex_desc *rx_desc;
                struct ice_rx_buf *rx_buf;
-               struct sk_buff *skb;
                unsigned int size;
                u16 stat_err_bits;
                u16 vlan_tag = 0;
@@ -1123,7 +1116,7 @@ int ice_clean_rx_irq(struct ice_ring *rx_ring, int budget)
                        ICE_RX_FLX_DESC_PKT_LEN_M;
 
                /* retrieve a buffer from the ring */
-               rx_buf = ice_get_rx_buf(rx_ring, &skb, size);
+               rx_buf = ice_get_rx_buf(rx_ring, size);
 
                if (!size) {
                        xdp.data = NULL;
@@ -1186,7 +1179,7 @@ int ice_clean_rx_irq(struct ice_ring *rx_ring, int budget)
                cleaned_count++;
 
                /* skip if it is NOP desc */
-               if (ice_is_non_eop(rx_ring, rx_desc, skb))
+               if (ice_is_non_eop(rx_ring, rx_desc))
                        continue;
 
                stat_err_bits = BIT(ICE_RX_FLEX_DESC_STATUS0_RXE_S);
@@ -1216,6 +1209,7 @@ int ice_clean_rx_irq(struct ice_ring *rx_ring, int budget)
 
                /* send completed skb up the stack */
                ice_receive_skb(rx_ring, skb, vlan_tag);
+               skb = NULL;
 
                /* update budget accounting */
                total_rx_pkts++;
@@ -1226,6 +1220,7 @@ int ice_clean_rx_irq(struct ice_ring *rx_ring, int budget)
 
        if (xdp_prog)
                ice_finalize_xdp_rx(rx_ring, xdp_xmit);
+       rx_ring->skb = skb;
 
        ice_update_rx_ring_stats(rx_ring, total_rx_pkts, total_rx_bytes);
 
diff --git a/drivers/net/ethernet/intel/ice/ice_txrx.h 
b/drivers/net/ethernet/intel/ice/ice_txrx.h
index ff1a1cbd078e..c77dbbb760cd 100644
--- a/drivers/net/ethernet/intel/ice/ice_txrx.h
+++ b/drivers/net/ethernet/intel/ice/ice_txrx.h
@@ -165,7 +165,6 @@ struct ice_tx_offload_params {
 struct ice_rx_buf {
        union {
                struct {
-                       struct sk_buff *skb;
                        dma_addr_t dma;
                        struct page *page;
                        unsigned int page_offset;
@@ -298,6 +297,7 @@ struct ice_ring {
        struct xsk_buff_pool *xsk_pool;
        /* CL3 - 3rd cacheline starts here */
        struct xdp_rxq_info xdp_rxq;
+       struct sk_buff *skb;
        /* CLX - the below items are only accessed infrequently and should be
         * in their own cache line if possible
         */
-- 
2.20.1

Reply via email to