03/12/2019 22:11, Jerome Tollet (jtollet):
> Thomas,
> I am afraid you may be missing the point. VPP is a framework where plugins 
> are first class citizens. If a plugin requires leveraging offload (inline or 
> lookaside), it is more than welcome to do it.
> There are multiple examples including hw crypto accelerators 
> (https://software.intel.com/en-us/articles/get-started-with-ipsec-acceleration-in-the-fdio-vpp-project).

OK I understand plugins are open.
My point was about the efficiency of the plugins,
given the need for buffer conversion.
If some plugins are already efficient enough, great:
it means there is no need for bringing effort in native VPP drivers.


> Le 03/12/2019 17:07, « vpp-dev@lists.fd.io au nom de Thomas Monjalon » 
> <vpp-dev@lists.fd.io au nom de tho...@monjalon.net> a écrit :
> 
>     03/12/2019 13:12, Damjan Marion:
>     > > On 3 Dec 2019, at 09:28, Thomas Monjalon <tho...@monjalon.net> wrote:
>     > > 03/12/2019 00:26, Damjan Marion:
>     > >> 
>     > >> Hi THomas!
>     > >> 
>     > >> Inline...
>     > >> 
>     > >>>> On 2 Dec 2019, at 23:35, Thomas Monjalon <tho...@monjalon.net> 
> wrote:
>     > >>> 
>     > >>> Hi all,
>     > >>> 
>     > >>> VPP has a buffer called vlib_buffer_t, while DPDK has rte_mbuf.
>     > >>> Are there some benchmarks about the cost of converting, from one 
> format
>     > >>> to the other one, during Rx/Tx operations?
>     > >> 
>     > >> We are benchmarking both dpdk i40e PMD performance and native VPP 
> AVF driver performance and we are seeing significantly better performance 
> with native AVF.
>     > >> If you taake a look at [1] you will see that DPDK i40e driver 
> provides 18.62 Mpps and exactly the same test with native AVF driver is 
> giving us arounf 24.86 Mpps.
>     [...]
>     > >> 
>     > >>> So why not improving DPDK integration in VPP to make it faster?
>     > >> 
>     > >> Yes, if we can get freedom to use parts of DPDK we want instead of 
> being forced to adopt whole DPDK ecosystem.
>     > >> for example, you cannot use dpdk drivers without using EAL, mempool, 
> rte_mbuf... rte_eal_init is monster which I was hoping that it will disappear 
> for long time...
>     > > 
>     > > You could help to improve these parts of DPDK,
>     > > instead of spending time to try implementing few drivers.
>     > > Then VPP would benefit from a rich driver ecosystem.
>     > 
>     > Thank you for letting me know what could be better use of my time.
>     
>     "You" was referring to VPP developers.
>     I think some other Cisco developers are also contributing to VPP.
>     
>     > At the moment we have good coverage of native drivers, and still there 
> is a option for people to use dpdk. It is now mainly up to driver vendors to 
> decide if they are happy with performance they wil get from dpdk pmd or they 
> want better...
>     
>     Yes it is possible to use DPDK in VPP with degraded performance.
>     If an user wants best performance with VPP and a real NIC,
>     a new driver must be implemented for VPP only.
>     
>     Anyway real performance benefits are in hardware device offloads
>     which will be hard to implement in VPP native drivers.
>     Support (investment) would be needed from vendors to make it happen.
>     About offloads, VPP is not using crypto or compression drivers
>     that DPDK provides (plus regex coming).
>     
>     VPP is a CPU-based packet processing software.
>     If users want to leverage hardware device offloads,
>     a truly DPDK-based software is required.
>     If I understand well your replies, such software cannot be VPP.



-=-=-=-=-=-=-=-=-=-=-=-
Links: You receive all messages sent to this group.

View/Reply Online (#14793): https://lists.fd.io/g/vpp-dev/message/14793
Mute This Topic: https://lists.fd.io/mt/65218320/21656
Group Owner: vpp-dev+ow...@lists.fd.io
Unsubscribe: https://lists.fd.io/g/vpp-dev/unsub  [arch...@mail-archive.com]
-=-=-=-=-=-=-=-=-=-=-=-

Reply via email to