> -----Original Message-----
> From: Jerin Jacob <jer...@marvell.com>
> Sent: Wednesday, January 29, 2025 1:13 PM
> To: Naga Harish K, S V <s.v.naga.haris...@intel.com>; Shijith Thotton
> <sthot...@marvell.com>; dev@dpdk.org
> Cc: Pavan Nikhilesh Bhagavatula <pbhagavat...@marvell.com>; Pathak,
> Pravin <pravin.pat...@intel.com>; Hemant Agrawal
> <hemant.agra...@nxp.com>; Sachin Saxena <sachin.sax...@nxp.com>;
> Mattias R_nnblom <mattias.ronnb...@ericsson.com>; Liang Ma
> <lian...@liangbit.com>; Mccarthy, Peter <peter.mccar...@intel.com>; Van
> Haaren, Harry <harry.van.haa...@intel.com>; Carrillo, Erik G
> <erik.g.carri...@intel.com>; Gujjar, Abhinandan S
> <abhinandan.guj...@intel.com>; Amit Prakash Shukla
> <amitpraka...@marvell.com>; Burakov, Anatoly
> <anatoly.bura...@intel.com>
> Subject: RE: [RFC PATCH] eventdev: adapter API to configure multiple Rx
> queues
> 
> 
> 
> > -----Original Message-----
> > From: Naga Harish K, S V <s.v.naga.haris...@intel.com>
> > Sent: Wednesday, January 29, 2025 10:35 AM
> > To: Shijith Thotton <sthot...@marvell.com>; dev@dpdk.org
> > Cc: Pavan Nikhilesh Bhagavatula <pbhagavat...@marvell.com>; Pathak,
> > Pravin <pravin.pat...@intel.com>; Hemant Agrawal
> > <hemant.agra...@nxp.com>; Sachin Saxena <sachin.sax...@nxp.com>;
> > Mattias R_nnblom <mattias.ronnb...@ericsson.com>; Jerin Jacob
> > <jer...@marvell.com>; Liang Ma <lian...@liangbit.com>; Mccarthy, Peter
> > <peter.mccar...@intel.com>; Van Haaren, Harry
> > <harry.van.haa...@intel.com>; Carrillo, Erik G
> > <erik.g.carri...@intel.com>; Gujjar, Abhinandan S
> > <abhinandan.guj...@intel.com>; Amit Prakash Shukla
> > <amitpraka...@marvell.com>; Burakov, Anatoly
> > <anatoly.bura...@intel.com>
> > Subject: [EXTERNAL] RE: [RFC PATCH] eventdev: adapter API to configure
> > multiple Rx queues
> > > >
> > > >This requires a change to the rte_event_eth_rx_adapter_queue_add()
> > > >stable API parameters.
> > > >This is an ABI breakage and may not be possible now.
> > > >It requires changes to many current applications that are using the
> > > >rte_event_eth_rx_adapter_queue_add() stable API.
> > > >
> > >
> > > What I meant by mapping was to retain the stable API parameters as they
> are.
> > > Internally, the API can use the proposed eventdev PMD operation
> > > (eth_rx_adapter_queues_add) without causing an ABI break, as shown
> below.
> > >
> > > int rte_event_eth_rx_adapter_queue_add(uint8_t id, uint16_t eth_dev_id,
> > >                 int32_t rx_queue_id,
> > >                 const struct rte_event_eth_rx_adapter_queue_conf *conf) {
> > >         if (rx_queue_id == -1)
> > >                 dev->dev_ops->eth_rx_adapter_queues_add)(
> > >                         dev, &rte_eth_devices[eth_dev_id], 0,
> > >                         conf, 0);
> > >         else
> > >                 dev->dev_ops->eth_rx_adapter_queues_add)(
> > >                         dev, &rte_eth_devices[eth_dev_id], &rx_queue_id,
> > >                         conf, 1);
> > > }
> > >
> > > With above change, old op (eth_rx_adapter_queue_add) can be removed
> > > as both API (stable and proposed) will be using
> eth_rx_adapter_queues_add.
> 
> 
> Since this thread is not converging and looks like it is due to confusion.
> I am trying to summarize my understanding to define the next steps(like if
> needed, we need to reach tech board if there are no consensus)
> 
> 
> Problem statement:
> ==================
> 1) Implementation of rte_event_eth_rx_adapter_queue_add() in HW typically
> uses an administrative function to enable it. Typically, it translated to 
> sending a
> mailbox to PF driver etc.
> So, this function takes "time" to complete in HW implementations.
> 2) For SW implementations, this won't take time as there is no other actors
> involved.
> 3) There are customer use cases, they add 300+
> rte_event_eth_rx_adapter_queue_add() on application bootup, that is
> introducing significant boot time for the application.
> Number of queues are function of number of ethdev ports, number  of
> ethdev Rx queues per port and number of event queues.
> 
> 
> Expected outcome of problem statement:
> ======================================
> 1) The cases where application knows queue mapping(typically at boot time
> case),
> application can call burst variant of rte_event_eth_rx_adapter_queue_add()
> function
> to amortize the cost. Similar scheme used DPDK in control path API where
> latency is critical,
> like rte_acl_add_rules() or rte_flow via template scheme.
> 2) Solution should not break ABI or any impact to SW drivers.
> 3) Avoid duplicating the code as much as possible
> 
> 
> Proposed solution:
> ==================
> 1) Update eventdev_eth_rx_adapter_queue_add_t() PMD (Internal ABI) API
> to take burst parameters
> 2) Add new rte_event_eth_rx_adapter_queue*s*_add() function and wire to
> use updated PMD API
> 3) Use rte_event_eth_rx_adapter_queue_add() as
> rte_event_eth_rx_adapter_queue*s*_add(...., 1)
> 
> If so, I am not sure what is the cons of this approach, it will let to have
> optimized applications when
> a) Application knows the queue mapping at priorly (typically in boot time)
> b) Allow HW drivers to optimize without breaking anything for SW drivers
> c) Provide applications to decide burst vs non burst selection based on the
> needed and performance requirements

The proposed API benefits only some hardware platforms that have optimized the 
"queue_add" eventdev PMD implementation for burst mode.
It may not benefit SW drivers/other HW platforms.
There will not be much difference in calling the existing API 
(rte_event_eth_rx_adapter_queue_add()) in a loop vs using the new API for the 
above cases.

If the new proposed API benefits all platforms, then it is useful.
This is the point I am making from the beginning, it is not captured in the 
summary.

Reply via email to