> On Jul 22, 2015, at 2:36 AM, Gray, Mark D <mark.d.gray at intel.com> wrote:
> 
> 
>> 
>> I'd like to hope that's my methodology problem, but I just follow the
>> installation guide without any customization.
>> 
>> Hi Mark, do you have any performance data share with us? Maybe we are
>> using different type of workloads, like I mentioned I am using typical data
>> center workload, I guess you are talking about NFV type of workload?
> 
> The number getting floated around on the mailing list recently is 16.5Mpps
> for phy-phy. However, I don't think we have any iperf data off-hand for your
> usecase. When we test throughput into the vm we usually generate the traffic 
> externally
> and send NIC->OVS->VM->OVS->NIC. This is a little different to your setup.
> 

I guess pmd driver is used inside VM in that case, right?
> I do know, however, that ovs-dpdk typically has a much larger throughput than
> the kernel space datapath.
> 

I'd like to say it depends on workloads, for small/medium packet size workload, 
that's definitely true, while for TSO size workload, it's not that obvious (or 
worse) as data path overheads are amortized and H/W can be leveraged.
> Have you seen this? 
> https://wiki.opnfv.org/characterize_vswitch_performance_for_telco_nfv_use_cases
> 

Thanks for the pointer, I'll try later.
>> 
>> Thanks,
>> Jun

Reply via email to