From: Alexander Lobakin <[email protected]>

[ Upstream commit bd74a86bc75d35adefbebcec7c3a743d02c06230 ]

Add the missing linking of NAPIs to netdev queues when enabling
interrupt vectors in order to support NAPI configuration and
interfaces requiring get_rx_queue()->napi to be set (like XSk
busy polling).

As currently, idpf_vport_{start,stop}() is called from several flows
with inconsistent RTNL locking, we need to synchronize them to avoid
runtime assertions. Notably:

* idpf_{open,stop}() -- regular NDOs, RTNL is always taken;
* idpf_initiate_soft_reset() -- usually called under RTNL;
* idpf_init_task -- called from the init work, needs RTNL;
* idpf_vport_dealloc -- called without RTNL taken, needs it.

Expand common idpf_vport_{start,stop}() to take an additional bool
telling whether we need to manually take the RTNL lock.

Suggested-by: Maciej Fijalkowski <[email protected]> # helper
Signed-off-by: Alexander Lobakin <[email protected]>
Tested-by: Ramu R <[email protected]>
Signed-off-by: Tony Nguyen <[email protected]>
Signed-off-by: Sasha Levin <[email protected]>
---

LLM Generated explanations, may be completely bogus:

YES

Why this is a bug-fix
- The driver was missing the association between netdev queues and their
  NAPI instances. This breaks NAPI-aware configuration and features that
  require queue->napi to be set, e.g., AF_XDP busy polling. The patch
  adds the missing linkage and corresponding unlinkage, which is clearly
  a functional fix rather than a feature.

What changed
- Link/unlink netdev queues to the NAPI of each q_vector:
  - Adds `idpf_q_vector_set_napi()` and uses it to associate both RX and
    TX queues with the q_vector’s `napi`:
    - Link on IRQ request:
      drivers/net/ethernet/intel/idpf/idpf_txrx.c:4043
    - Unlink on IRQ free:
      drivers/net/ethernet/intel/idpf/idpf_txrx.c:3852
  - Helper implementation:
    drivers/net/ethernet/intel/idpf/idpf_txrx.c:3818

- Ensure correct locking for netif_queue_set_napi:
  - `netif_queue_set_napi()` asserts RTNL or invisibility
    (net/core/dev.c:7167), so the patch adds an `rtnl` parameter to the
    vport bring-up/tear-down paths and acquires RTNL where it previously
    wasn’t guaranteed:
    - `idpf_vport_open(struct idpf_vport *vport, bool rtnl)` acquires
      RTNL when `rtnl=true`
      (drivers/net/ethernet/intel/idpf/idpf_lib.c:1397–1400), and
      releases on both success and error paths (1528–1531).
    - `idpf_vport_stop(struct idpf_vport *vport, bool rtnl)` does the
      same for teardown (900–927).
  - Callers updated according to their RTNL context, avoiding double-
    lock or missing-lock situations:
    - NDO stop: passes `false` (called under RTNL):
      drivers/net/ethernet/intel/idpf/idpf_lib.c:951
    - NDO open: passes `false` (called under RTNL):
      drivers/net/ethernet/intel/idpf/idpf_lib.c:2275
    - init work (not under RTNL): `idpf_init_task()` passes `true`:
      drivers/net/ethernet/intel/idpf/idpf_lib.c:1607
    - vport dealloc (not under RTNL): passes `true`:
      drivers/net/ethernet/intel/idpf/idpf_lib.c:1044
    - soft reset (usually under RTNL via ndo contexts): passes `false`:
      drivers/net/ethernet/intel/idpf/idpf_lib.c:1997 and reopen at
      2027, 2037

- Order of operations remains sane:
  - Add NAPI and map vectors, then request IRQs, then link queues to
    NAPI, then enable NAPI/IRQs
    (drivers/net/ethernet/intel/idpf/idpf_txrx.c:4598–4607, 4043,
    4619–4621).
  - On teardown disable interrupts/NAPI, delete NAPI, unlink queues,
    free IRQs (drivers/net/ethernet/intel/idpf/idpf_txrx.c:4119–4125,
    3852).

Impact and risk
- User-visible bug fixed: AF_XDP busy-polling and other NAPI-aware paths
  can now retrieve the correct NAPI via get_rx_queue()->napi.
- Change is tightly scoped to the idpf driver; no UAPI or architectural
  changes.
- Locking adjustments are minimal and consistent with net core
  expectations for `netif_queue_set_napi()`.
- Similar pattern exists in other drivers (e.g., ice, igb, igc) that use
  `netif_queue_set_napi`, which supports the approach’s correctness.
- Note: In the rare request_irq failure unwind, the code frees any
  requested IRQs but doesn’t explicitly clear queue->napi for
  previously-linked vectors; however, `napi_del()` runs and the
  q_vector/napi storage remains valid, and normal teardown does clear
  associations. This is a minor edge and does not outweigh the benefit
  of the fix.

Stable backport suitability
- Meets stable criteria: fixes a real functional bug, small and self-
  contained, limited to a single driver, low regression risk, and
  conforms to net core locking rules.
- Dependency: requires `netif_queue_set_napi()` (present in this branch,
  net/core/dev.c:7159). For older stable series lacking this API, a
  backport would need equivalent infrastructure or adaptation.

Conclusion
- This is a clear, necessary bug fix enabling expected NAPI-aware
  behavior in idpf. It is safe and appropriate to backport.

 drivers/net/ethernet/intel/idpf/idpf_lib.c  | 38 +++++++++++++++------
 drivers/net/ethernet/intel/idpf/idpf_txrx.c | 17 +++++++++
 2 files changed, 45 insertions(+), 10 deletions(-)

diff --git a/drivers/net/ethernet/intel/idpf/idpf_lib.c 
b/drivers/net/ethernet/intel/idpf/idpf_lib.c
index e327950c93d8e..f4b89d222610f 100644
--- a/drivers/net/ethernet/intel/idpf/idpf_lib.c
+++ b/drivers/net/ethernet/intel/idpf/idpf_lib.c
@@ -884,14 +884,18 @@ static void idpf_remove_features(struct idpf_vport *vport)
 /**
  * idpf_vport_stop - Disable a vport
  * @vport: vport to disable
+ * @rtnl: whether to take RTNL lock
  */
-static void idpf_vport_stop(struct idpf_vport *vport)
+static void idpf_vport_stop(struct idpf_vport *vport, bool rtnl)
 {
        struct idpf_netdev_priv *np = netdev_priv(vport->netdev);
 
        if (np->state <= __IDPF_VPORT_DOWN)
                return;
 
+       if (rtnl)
+               rtnl_lock();
+
        netif_carrier_off(vport->netdev);
        netif_tx_disable(vport->netdev);
 
@@ -913,6 +917,9 @@ static void idpf_vport_stop(struct idpf_vport *vport)
        idpf_vport_queues_rel(vport);
        idpf_vport_intr_rel(vport);
        np->state = __IDPF_VPORT_DOWN;
+
+       if (rtnl)
+               rtnl_unlock();
 }
 
 /**
@@ -936,7 +943,7 @@ static int idpf_stop(struct net_device *netdev)
        idpf_vport_ctrl_lock(netdev);
        vport = idpf_netdev_to_vport(netdev);
 
-       idpf_vport_stop(vport);
+       idpf_vport_stop(vport, false);
 
        idpf_vport_ctrl_unlock(netdev);
 
@@ -1029,7 +1036,7 @@ static void idpf_vport_dealloc(struct idpf_vport *vport)
        idpf_idc_deinit_vport_aux_device(vport->vdev_info);
 
        idpf_deinit_mac_addr(vport);
-       idpf_vport_stop(vport);
+       idpf_vport_stop(vport, true);
 
        if (!test_bit(IDPF_HR_RESET_IN_PROG, adapter->flags))
                idpf_decfg_netdev(vport);
@@ -1370,8 +1377,9 @@ static void idpf_rx_init_buf_tail(struct idpf_vport 
*vport)
 /**
  * idpf_vport_open - Bring up a vport
  * @vport: vport to bring up
+ * @rtnl: whether to take RTNL lock
  */
-static int idpf_vport_open(struct idpf_vport *vport)
+static int idpf_vport_open(struct idpf_vport *vport, bool rtnl)
 {
        struct idpf_netdev_priv *np = netdev_priv(vport->netdev);
        struct idpf_adapter *adapter = vport->adapter;
@@ -1381,6 +1389,9 @@ static int idpf_vport_open(struct idpf_vport *vport)
        if (np->state != __IDPF_VPORT_DOWN)
                return -EBUSY;
 
+       if (rtnl)
+               rtnl_lock();
+
        /* we do not allow interface up just yet */
        netif_carrier_off(vport->netdev);
 
@@ -1388,7 +1399,7 @@ static int idpf_vport_open(struct idpf_vport *vport)
        if (err) {
                dev_err(&adapter->pdev->dev, "Failed to allocate interrupts for 
vport %u: %d\n",
                        vport->vport_id, err);
-               return err;
+               goto err_rtnl_unlock;
        }
 
        err = idpf_vport_queues_alloc(vport);
@@ -1475,6 +1486,9 @@ static int idpf_vport_open(struct idpf_vport *vport)
                goto deinit_rss;
        }
 
+       if (rtnl)
+               rtnl_unlock();
+
        return 0;
 
 deinit_rss:
@@ -1492,6 +1506,10 @@ static int idpf_vport_open(struct idpf_vport *vport)
 intr_rel:
        idpf_vport_intr_rel(vport);
 
+err_rtnl_unlock:
+       if (rtnl)
+               rtnl_unlock();
+
        return err;
 }
 
@@ -1572,7 +1590,7 @@ void idpf_init_task(struct work_struct *work)
        np = netdev_priv(vport->netdev);
        np->state = __IDPF_VPORT_DOWN;
        if (test_and_clear_bit(IDPF_VPORT_UP_REQUESTED, vport_config->flags))
-               idpf_vport_open(vport);
+               idpf_vport_open(vport, true);
 
        /* Spawn and return 'idpf_init_task' work queue until all the
         * default vports are created
@@ -1962,7 +1980,7 @@ int idpf_initiate_soft_reset(struct idpf_vport *vport,
                idpf_send_delete_queues_msg(vport);
        } else {
                set_bit(IDPF_VPORT_DEL_QUEUES, vport->flags);
-               idpf_vport_stop(vport);
+               idpf_vport_stop(vport, false);
        }
 
        idpf_deinit_rss(vport);
@@ -1992,7 +2010,7 @@ int idpf_initiate_soft_reset(struct idpf_vport *vport,
                goto err_open;
 
        if (current_state == __IDPF_VPORT_UP)
-               err = idpf_vport_open(vport);
+               err = idpf_vport_open(vport, false);
 
        goto free_vport;
 
@@ -2002,7 +2020,7 @@ int idpf_initiate_soft_reset(struct idpf_vport *vport,
 
 err_open:
        if (current_state == __IDPF_VPORT_UP)
-               idpf_vport_open(vport);
+               idpf_vport_open(vport, false);
 
 free_vport:
        kfree(new_vport);
@@ -2240,7 +2258,7 @@ static int idpf_open(struct net_device *netdev)
        if (err)
                goto unlock;
 
-       err = idpf_vport_open(vport);
+       err = idpf_vport_open(vport, false);
 
 unlock:
        idpf_vport_ctrl_unlock(netdev);
diff --git a/drivers/net/ethernet/intel/idpf/idpf_txrx.c 
b/drivers/net/ethernet/intel/idpf/idpf_txrx.c
index e75a94d7ac2ac..92634c4bb369a 100644
--- a/drivers/net/ethernet/intel/idpf/idpf_txrx.c
+++ b/drivers/net/ethernet/intel/idpf/idpf_txrx.c
@@ -3430,6 +3430,20 @@ void idpf_vport_intr_rel(struct idpf_vport *vport)
        vport->q_vectors = NULL;
 }
 
+static void idpf_q_vector_set_napi(struct idpf_q_vector *q_vector, bool link)
+{
+       struct napi_struct *napi = link ? &q_vector->napi : NULL;
+       struct net_device *dev = q_vector->vport->netdev;
+
+       for (u32 i = 0; i < q_vector->num_rxq; i++)
+               netif_queue_set_napi(dev, q_vector->rx[i]->idx,
+                                    NETDEV_QUEUE_TYPE_RX, napi);
+
+       for (u32 i = 0; i < q_vector->num_txq; i++)
+               netif_queue_set_napi(dev, q_vector->tx[i]->idx,
+                                    NETDEV_QUEUE_TYPE_TX, napi);
+}
+
 /**
  * idpf_vport_intr_rel_irq - Free the IRQ association with the OS
  * @vport: main vport structure
@@ -3450,6 +3464,7 @@ static void idpf_vport_intr_rel_irq(struct idpf_vport 
*vport)
                vidx = vport->q_vector_idxs[vector];
                irq_num = adapter->msix_entries[vidx].vector;
 
+               idpf_q_vector_set_napi(q_vector, false);
                kfree(free_irq(irq_num, q_vector));
        }
 }
@@ -3637,6 +3652,8 @@ static int idpf_vport_intr_req_irq(struct idpf_vport 
*vport)
                                   "Request_irq failed, error: %d\n", err);
                        goto free_q_irqs;
                }
+
+               idpf_q_vector_set_napi(q_vector, true);
        }
 
        return 0;
-- 
2.51.0

Reply via email to