Thank you Ilya.
This is the output of coverage/show during VM port configuration:
# ovs-appctl coverage/show
Event coverage, avg rate over last: 5 seconds, last minute, last hour,
hash=e0094f87:
bridge_reconfigure 0.0/sec 1.017/sec 0.0169/sec total: 739
ofproto_flush 0.0/sec 0.000/sec 0.0000/sec total: 4
ofproto_packet_out 0.0/sec 0.183/sec 0.1158/sec total: 7195
ofproto_recv_openflow 642.0/sec 669.883/sec 16.7508/sec
total: 1557388
ofproto_update_port 0.0/sec 0.900/sec 0.0150/sec total: 470
ofproto_dpif_expired 0.0/sec 0.000/sec 0.0000/sec total: 1
rev_reconfigure 0.0/sec 1.017/sec 0.0181/sec total: 872
rev_bond 0.0/sec 0.000/sec 0.0000/sec total: 3
rev_port_toggled 0.0/sec 0.000/sec 0.0000/sec total: 58
rev_flow_table 321.0/sec 330.500/sec 7.9622/sec
total: 752381
rev_mac_learning 0.2/sec 0.033/sec 0.0556/sec total: 3018
xlate_actions 2136.6/sec 2681.233/sec 65.8039/sec
total: 7553425
ccmap_expand 0.0/sec 0.000/sec 0.0000/sec total: 997
ccmap_shrink 0.4/sec 0.333/sec 0.2906/sec
total: 18702
cmap_expand 0.6/sec 0.500/sec 0.5500/sec
total: 34155
cmap_shrink 0.4/sec 0.417/sec 0.4608/sec
total: 26461
datapath_drop_upcall_error 0.0/sec 0.000/sec 0.0000/sec total: 6
datapath_drop_userspace_action_error 0.0/sec 0.333/sec
0.2633/sec total: 15866
dpif_port_add 0.0/sec 0.450/sec 0.0075/sec total: 112
dpif_port_del 0.0/sec 0.000/sec 0.0000/sec total: 152
dpif_flow_flush 0.0/sec 0.000/sec 0.0000/sec total: 5
dpif_flow_get 0.0/sec 0.000/sec 0.0000/sec total: 24
dpif_flow_put 0.0/sec 0.000/sec 0.0000/sec total: 293
dpif_flow_del 1.6/sec 1.367/sec 1.4736/sec
total: 85778
dpif_execute 0.0/sec 0.250/sec 0.1825/sec
total: 120001
dpif_meter_set 0.0/sec 0.000/sec 0.0000/sec total: 16
flow_extract 0.0/sec 0.400/sec 0.3300/sec
total: 128674
miniflow_malloc 642.0/sec 666.000/sec 16.5678/sec
total: 1550142
hindex_expand 0.0/sec 0.000/sec 0.0000/sec total: 11
hmap_pathological 1284.8/sec 971.267/sec 19.9536/sec
total: 1807500
hmap_expand 11104.8/sec 11043.833/sec 285.0853/sec
total: 26064158
hmap_shrink 0.0/sec 13.250/sec 0.2208/sec total: 5216
mac_learning_learned 0.2/sec 0.017/sec 0.0481/sec total: 2705
mac_learning_expired 0.0/sec 0.017/sec 0.0492/sec total: 2685
netdev_received 3.2/sec 2.083/sec 2.6144/sec
total: 167877
netdev_sent 0.2/sec 0.400/sec 0.4772/sec
total: 137686
netdev_get_stats 17.2/sec 14.983/sec 11.7547/sec
total: 689284
drop_action_of_pipeline 3.0/sec 1.733/sec 2.2069/sec
total: 144682
txn_unchanged 0.0/sec 0.467/sec 0.0733/sec total: 4049
txn_incomplete 0.4/sec 2.083/sec 0.2419/sec
total: 13933
txn_success 0.2/sec 0.800/sec 0.2083/sec
total: 12025
txn_try_again 0.0/sec 0.000/sec 0.0000/sec total: 8
poll_create_node 76372.0/sec 79107.450/sec 2382.4131/sec
total: 219988137
poll_zero_timeout 2721.6/sec 2843.933/sec 88.4772/sec
total: 8713208
rconn_queued 321.0/sec 336.517/sec 8.8169/sec
total: 807033
rconn_sent 321.0/sec 336.517/sec 8.8169/sec
total: 807033
seq_change 137170.8/sec 140900.517/sec 50769.3464/sec
total: 3189942824
pstream_open 0.0/sec 0.000/sec 0.0000/sec total: 23
stream_open 0.0/sec 0.000/sec 0.0000/sec total: 21
unixctl_received 0.0/sec 0.033/sec 0.0014/sec total: 5
unixctl_replied 0.0/sec 0.033/sec 0.0014/sec total: 5
util_xalloc 296344.6/sec 288438.433/sec 8485.0417/sec
total: 764536339
vconn_open 0.0/sec 0.000/sec 0.0000/sec total: 20
vconn_received 642.0/sec 672.600/sec 16.7978/sec
total: 1558795
vconn_sent 321.0/sec 339.233/sec 8.8639/sec
total: 809031
netdev_set_policing 0.0/sec 0.000/sec 0.0000/sec total: 40
netdev_arp_lookup 0.0/sec 0.000/sec 0.0000/sec total: 8
netdev_get_ifindex 0.0/sec 0.000/sec 0.0000/sec total: 56
netdev_set_hwaddr 0.0/sec 0.000/sec 0.0000/sec total: 4
netdev_get_ethtool 0.0/sec 0.000/sec 0.0000/sec total: 56
netlink_received 2.0/sec 31.667/sec 3.1333/sec
total: 174488
netlink_recv_jumbo 2.0/sec 31.667/sec 3.1333/sec
total: 171692
netlink_sent 2.0/sec 31.667/sec 3.1333/sec
total: 174276
vhost_notification 0.0/sec 0.067/sec 0.0047/sec total: 895
nln_changed 0.0/sec 0.000/sec 0.0000/sec total: 74
So does this line "ofproto_recv_openflow 642.0/sec 669.883/sec
16.7508/sec total: 1557388"
mean that the controller is a bit slow to call OPENFLOW?
What should be the baseline performance here?
This is the coverage/show before creating VM (install flows):
# ovs-appctl coverage/show
Event coverage, avg rate over last: 5 seconds, last minute, last hour,
hash=90443f93:
bridge_reconfigure 0.0/sec 0.000/sec 0.0000/sec total: 678
ofproto_flush 0.0/sec 0.000/sec 0.0000/sec total: 4
ofproto_packet_out 0.0/sec 0.117/sec 0.1158/sec total: 7170
ofproto_recv_openflow 0.4/sec 4.000/sec 3.9231/sec
total: 1508269
ofproto_update_port 0.0/sec 0.000/sec 0.0000/sec total: 416
ofproto_dpif_expired 0.0/sec 0.000/sec 0.0000/sec total: 1
rev_reconfigure 0.0/sec 0.000/sec 0.0011/sec total: 810
rev_bond 0.0/sec 0.000/sec 0.0000/sec total: 3
rev_port_toggled 0.0/sec 0.000/sec 0.0000/sec total: 58
rev_flow_table 0.0/sec 1.650/sec 1.6189/sec
total: 728140
rev_mac_learning 0.0/sec 0.083/sec 0.0572/sec total: 3011
xlate_actions 0.0/sec 17.050/sec 17.5286/sec
total: 7364896
ccmap_expand 0.0/sec 0.000/sec 0.0000/sec total: 997
ccmap_shrink 0.2/sec 0.250/sec 0.2906/sec
total: 18643
cmap_expand 0.0/sec 0.500/sec 0.5642/sec
total: 34071
cmap_shrink 0.4/sec 0.367/sec 0.4719/sec
total: 26389
datapath_drop_upcall_error 0.0/sec 0.000/sec 0.0000/sec total: 6
datapath_drop_userspace_action_error 0.0/sec 0.267/sec
0.2633/sec total: 15814
dpif_port_add 0.0/sec 0.000/sec 0.0000/sec total: 85
dpif_port_del 0.0/sec 0.000/sec 0.0000/sec total: 152
dpif_flow_flush 0.0/sec 0.000/sec 0.0000/sec total: 5
dpif_flow_get 0.0/sec 0.000/sec 0.0000/sec total: 24
dpif_flow_put 0.0/sec 0.000/sec 0.0000/sec total: 293
dpif_flow_del 4.8/sec 1.617/sec 1.4819/sec
total: 85505
dpif_execute 0.0/sec 0.183/sec 0.1819/sec
total: 119962
dpif_meter_set 0.0/sec 0.000/sec 0.0000/sec total: 16
flow_extract 0.0/sec 0.333/sec 0.3294/sec
total: 128608
miniflow_malloc 0.4/sec 3.867/sec 3.8039/sec
total: 1501278
hindex_expand 0.0/sec 0.000/sec 0.0000/sec total: 11
hmap_pathological 0.0/sec 2.600/sec 2.5167/sec
total: 1739823
hmap_expand 24.0/sec 78.450/sec 76.2433/sec
total: 25258980
hmap_shrink 0.0/sec 0.000/sec 0.0000/sec total: 4421
mac_learning_learned 0.0/sec 0.117/sec 0.0517/sec total: 2702
mac_learning_expired 0.0/sec 0.033/sec 0.0500/sec total: 2677
netdev_received 0.4/sec 4.517/sec 2.7444/sec
total: 167482
netdev_sent 0.0/sec 0.417/sec 0.4844/sec
total: 137611
netdev_get_stats 11.8/sec 11.800/sec 11.6197/sec
total: 686588
drop_action_of_pipeline 0.4/sec 4.367/sec 2.3283/sec
total: 144346
txn_unchanged 0.0/sec 0.067/sec 0.0661/sec total: 4011
txn_incomplete 0.2/sec 0.200/sec 0.2092/sec
total: 13773
txn_success 0.2/sec 0.200/sec 0.1969/sec
total: 11947
txn_try_again 0.0/sec 0.000/sec 0.0000/sec total: 8
poll_create_node 433.0/sec 882.017/sec 869.8861/sec
total: 214105896
poll_zero_timeout 16.2/sec 28.200/sec 34.0678/sec
total: 8499849
rconn_queued 0.4/sec 2.433/sec 2.3825/sec
total: 782322
rconn_sent 0.4/sec 2.433/sec 2.3825/sec
total: 782322
seq_change 49320.6/sec 49483.567/sec 48596.7650/sec
total: 3172852337
pstream_open 0.0/sec 0.000/sec 0.0000/sec total: 23
stream_open 0.0/sec 0.000/sec 0.0000/sec total: 21
unixctl_received 0.0/sec 0.033/sec 0.0006/sec total: 2
unixctl_replied 0.0/sec 0.033/sec 0.0006/sec total: 2
util_xalloc 1501.6/sec 3101.350/sec 3057.6908/sec
total: 743357369
vconn_open 0.0/sec 0.000/sec 0.0000/sec total: 20
vconn_received 0.4/sec 4.000/sec 3.9231/sec
total: 1509507
vconn_sent 0.4/sec 2.433/sec 2.3825/sec
total: 784151
netdev_set_policing 0.0/sec 0.000/sec 0.0000/sec total: 40
netdev_arp_lookup 0.0/sec 0.000/sec 0.0000/sec total: 8
netdev_get_ifindex 0.0/sec 0.000/sec 0.0000/sec total: 56
netdev_set_hwaddr 0.0/sec 0.000/sec 0.0000/sec total: 4
netdev_get_ethtool 0.0/sec 0.000/sec 0.0000/sec total: 56
netlink_received 2.0/sec 2.667/sec 2.6306/sec
total: 172188
netlink_recv_jumbo 2.0/sec 2.667/sec 2.6306/sec
total: 169392
netlink_sent 2.0/sec 2.667/sec 2.6306/sec
total: 171976
vhost_notification 0.0/sec 0.000/sec 0.0086/sec total: 891
nln_changed 0.0/sec 0.000/sec 0.0000/sec total: 74
The data plane currently has almost no traffic, so there are relatively
few upcall processes. The main operation is for the controller to install
the flow through OpenFlow.
Hope this email can serve as a guide for people facing similar problems
in the future.
How to identify system bottlenecks? How to optimize?
Regards,
LIU Yulong
On Mon, Sep 23, 2024 at 7:48 PM Ilya Maximets <[email protected]> wrote:
>
> On 9/23/24 12:22, LIU Yulong via discuss wrote:
> > Hi there,
> >
> > In an OVS-DPDK environment, we noticed that the ovs flow installation
> > is too slow to finish the process of our VM NIC configuration.
> >
> > Here are some logs [1], you may see some text like this:
> >
> > 20482 flow_mods in the last 59 s (20362 adds, 120 deletes)
> > 20265 flow_mods in the last 59 s (19888 adds, 377 deletes)
> >
> > When concurrent port creation comes, 59s for many ports are not so
> > efficient.
>
> Hi. In general, this log message doesn't mean that it took 59 seconds
> to perform 20K flow modifications. It means that in the last 59 seconds
> there were 20K flow modifications, i.e. the process might have slept most
> of that time waking up periodically to process some updates.
>
> And you may see that even when the process was busy waking up on RCU, it
> was not using 100% of CPU.
>
> In order to better track the rate of incoming flow modifications,
> you may look at coverage counters instead.
>
> Stats of the OVS threads also do nott indicate any high load on the
> main ovs-vswitchd thread that is responsible for the flow mods.
> Revalidator threads also do not seem to be particularly busy.
>
> Best regards, Ilya Maximets.
>
> >
> > So here I wonder what is the benchmark performance for OVS flow table
> > installation? How to perform performance tuning for flow installation?
> >
> > Ovs version is 2.17.2, and dpdk is 20.11.
> > ovs-vswitchd threads are [2].
> >
> >
> >
> >
> >
> > Regards,
> >
> > LIU Yulong
> >
> >
> >
> >
> >
> > [1] LOGs from ovs-vswitchd:
> > 2024-09-23T10:21:29.168Z|08389|poll_loop|INFO|Dropped 256 log messages
> > in last 27 seconds (most recently, 27 seconds ago) due to excessive
> > rate
> > 2024-09-23T10:21:29.168Z|08390|poll_loop|INFO|wakeup due to [POLLIN]
> > on fd 314 (FIFO pipe:[872871807]) at lib/ovs-rcu.c:249 (52% CPU usage)
> > 2024-09-23T10:21:29.168Z|08391|poll_loop|INFO|wakeup due to [POLLIN]
> > on fd 314 (FIFO pipe:[872871807]) at lib/ovs-rcu.c:249 (52% CPU usage)
> > 2024-09-23T10:21:29.168Z|08392|poll_loop|INFO|wakeup due to [POLLIN]
> > on fd 314 (FIFO pipe:[872871807]) at lib/ovs-rcu.c:249 (52% CPU usage)
> > 2024-09-23T10:21:29.168Z|08393|poll_loop|INFO|wakeup due to [POLLIN]
> > on fd 314 (FIFO pipe:[872871807]) at lib/ovs-rcu.c:249 (52% CPU usage)
> > 2024-09-23T10:21:32.136Z|08394|poll_loop|INFO|Dropped 40909 log
> > messages in last 3 seconds (most recently, 0 seconds ago) due to
> > excessive rate
> > 2024-09-23T10:21:32.136Z|08395|poll_loop|INFO|wakeup due to [POLLIN]
> > on fd 314 (FIFO pipe:[872871807]) at lib/ovs-rcu.c:249 (52% CPU usage)
> > 2024-09-23T10:22:08.981Z|08396|connmgr|INFO|br-int<->tcp:127.0.0.1:6633:
> > 20265 flow_mods in the last 59 s (19888 adds, 377 deletes)
> > 2024-09-23T10:22:14.820Z|08397|connmgr|INFO|br-dpdk<->tcp:127.0.0.1:6633:
> > 103 flow_mods 10 s ago (103 adds)
> > 2024-09-23T10:22:35.211Z|08398|poll_loop|INFO|Dropped 392 log messages
> > in last 63 seconds (most recently, 63 seconds ago) due to excessive
> > rate
> > 2024-09-23T10:22:35.211Z|08399|poll_loop|INFO|wakeup due to [POLLIN]
> > on fd 184 (127.0.0.1:55906<->127.0.0.1:6633) at lib/stream-fd.c:157
> > (57% CPU usage)
> > 2024-09-23T10:22:35.211Z|08400|poll_loop|INFO|wakeup due to [POLLIN]
> > on fd 314 (FIFO pipe:[872871807]) at lib/ovs-rcu.c:249 (57% CPU usage)
> > 2024-09-23T10:22:35.211Z|08401|poll_loop|INFO|wakeup due to [POLLIN]
> > on fd 314 (FIFO pipe:[872871807]) at lib/ovs-rcu.c:249 (57% CPU usage)
> > 2024-09-23T10:22:35.211Z|08402|poll_loop|INFO|wakeup due to [POLLIN]
> > on fd 314 (FIFO pipe:[872871807]) at lib/ovs-rcu.c:249 (57% CPU usage)
> > 2024-09-23T10:22:35.211Z|08403|poll_loop|INFO|wakeup due to [POLLIN]
> > on fd 314 (FIFO pipe:[872871807]) at lib/ovs-rcu.c:249 (57% CPU usage)
> > 2024-09-23T10:22:35.211Z|08404|poll_loop|INFO|wakeup due to [POLLIN]
> > on fd 314 (FIFO pipe:[872871807]) at lib/ovs-rcu.c:249 (57% CPU usage)
> > 2024-09-23T10:22:35.211Z|08405|poll_loop|INFO|wakeup due to [POLLIN]
> > on fd 314 (FIFO pipe:[872871807]) at lib/ovs-rcu.c:249 (57% CPU usage)
> > 2024-09-23T10:22:35.211Z|08406|poll_loop|INFO|wakeup due to [POLLIN]
> > on fd 314 (FIFO pipe:[872871807]) at lib/ovs-rcu.c:249 (57% CPU usage)
> > 2024-09-23T10:22:35.211Z|08407|poll_loop|INFO|wakeup due to [POLLIN]
> > on fd 314 (FIFO pipe:[872871807]) at lib/ovs-rcu.c:249 (57% CPU usage)
> > 2024-09-23T10:22:35.211Z|08408|poll_loop|INFO|wakeup due to [POLLIN]
> > on fd 314 (FIFO pipe:[872871807]) at lib/ovs-rcu.c:249 (57% CPU usage)
> > 2024-09-23T10:23:08.981Z|08409|connmgr|INFO|br-int<->tcp:127.0.0.1:6633:
> > 20112 flow_mods in the last 59 s (19860 adds, 252 deletes)
> > 2024-09-23T10:23:14.821Z|08410|connmgr|INFO|br-dpdk<->tcp:127.0.0.1:6633:
> > 103 flow_mods 10 s ago (103 adds)
> > 2024-09-23T10:23:44.223Z|08411|poll_loop|INFO|Dropped 35966 log
> > messages in last 69 seconds (most recently, 66 seconds ago) due to
> > excessive rate
> > 2024-09-23T10:23:44.224Z|08412|poll_loop|INFO|wakeup due to [POLLIN]
> > on fd 184 (127.0.0.1:55906<->127.0.0.1:6633) at lib/stream-fd.c:157
> > (81% CPU usage)
> > 2024-09-23T10:23:44.224Z|08413|poll_loop|INFO|wakeup due to [POLLIN]
> > on fd 314 (FIFO pipe:[872871807]) at lib/ovs-rcu.c:249 (81% CPU usage)
> > 2024-09-23T10:23:44.224Z|08414|poll_loop|INFO|wakeup due to [POLLIN]
> > on fd 314 (FIFO pipe:[872871807]) at lib/ovs-rcu.c:249 (81% CPU usage)
> > 2024-09-23T10:23:44.224Z|08415|poll_loop|INFO|wakeup due to [POLLIN]
> > on fd 314 (FIFO pipe:[872871807]) at lib/ovs-rcu.c:249 (81% CPU usage)
> > 2024-09-23T10:23:44.224Z|08416|poll_loop|INFO|wakeup due to [POLLIN]
> > on fd 314 (FIFO pipe:[872871807]) at lib/ovs-rcu.c:249 (81% CPU usage)
> > 2024-09-23T10:23:44.224Z|08417|poll_loop|INFO|wakeup due to [POLLIN]
> > on fd 314 (FIFO pipe:[872871807]) at lib/ovs-rcu.c:249 (81% CPU usage)
> > 2024-09-23T10:23:44.224Z|08418|poll_loop|INFO|wakeup due to [POLLIN]
> > on fd 314 (FIFO pipe:[872871807]) at lib/ovs-rcu.c:249 (81% CPU usage)
> > 2024-09-23T10:23:44.224Z|08419|poll_loop|INFO|wakeup due to 0-ms
> > timeout at lib/ovs-rcu.c:249 (81% CPU usage)
> > 2024-09-23T10:23:44.224Z|08420|poll_loop|INFO|wakeup due to [POLLIN]
> > on fd 314 (FIFO pipe:[872871807]) at lib/ovs-rcu.c:249 (81% CPU usage)
> > 2024-09-23T10:23:44.224Z|08421|poll_loop|INFO|wakeup due to [POLLIN]
> > on fd 314 (FIFO pipe:[872871807]) at lib/ovs-rcu.c:249 (81% CPU usage)
> > 2024-09-23T10:23:50.225Z|08422|poll_loop|INFO|Dropped 13989 log
> > messages in last 6 seconds (most recently, 0 seconds ago) due to
> > excessive rate
> > 2024-09-23T10:23:50.225Z|08423|poll_loop|INFO|wakeup due to 0-ms
> > timeout at ofproto/ofproto-dpif.c:2001 (88% CPU usage)
> > 2024-09-23T10:23:56.625Z|08424|poll_loop|INFO|Dropped 28118 log
> > messages in last 6 seconds (most recently, 0 seconds ago) due to
> > excessive rate
> > 2024-09-23T10:23:56.625Z|08425|poll_loop|INFO|wakeup due to [POLLIN]
> > on fd 314 (FIFO pipe:[872871807]) at vswitchd/bridge.c:423 (59% CPU
> > usage)
> > 2024-09-23T10:24:08.122Z|08426|connmgr|INFO|br-meta<->tcp:127.0.0.1:6633:
> > 108 flow_mods in the 7 s starting 10 s ago (108 deletes)
> > 2024-09-23T10:24:08.981Z|08427|connmgr|INFO|br-int<->tcp:127.0.0.1:6633:
> > 14552 flow_mods in the 56 s starting 59 s ago (11093 adds, 3459
> > deletes)
> > 2024-09-23T10:24:14.821Z|08428|connmgr|INFO|br-dpdk<->tcp:127.0.0.1:6633:
> > 589 flow_mods in the 7 s starting 16 s ago (103 adds, 486 deletes)
> > 2024-09-23T10:25:14.821Z|08429|connmgr|INFO|br-dpdk<->tcp:127.0.0.1:6633:
> > 103 flow_mods 10 s ago (103 adds)
> > 2024-09-23T10:26:14.821Z|08430|connmgr|INFO|br-dpdk<->tcp:127.0.0.1:6633:
> > 103 flow_mods 10 s ago (103 adds)
> >
> > [2] Threads of the ovs-vswitchd:
> > PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+
> > COMMAND
> > 13148 root 10 -10 241.2g 1.3g 33016 R 99.9 0.3 124:04.40
> > pmd-c37/id:20
> > 13156 root 10 -10 241.2g 1.3g 33016 R 80.0 0.3 124:06.10
> > pmd-c50/id:28
> > 13144 root 10 -10 241.2g 1.3g 33016 R 73.3 0.3 124:06.07
> > pmd-c53/id:16
> > 13151 root 10 -10 241.2g 1.3g 33016 R 73.3 0.3 124:06.09
> > pmd-c51/id:23
> > 13155 root 10 -10 241.2g 1.3g 33016 R 73.3 0.3 124:06.09
> > pmd-c52/id:27
> > 13111 root 10 -10 241.2g 1.3g 33016 S 0.0 0.3 9:38.57
> > ovs-vswitchd
> > 13112 root 10 -10 241.2g 1.3g 33016 S 0.0 0.3 0:00.00
> > eal-intr-thread
> > 13113 root 10 -10 241.2g 1.3g 33016 S 0.0 0.3 0:00.00
> > rte_mp_handle
> > 13118 root 10 -10 241.2g 1.3g 33016 S 0.0 0.3 0:00.00
> > ovs-vswitchd
> > 13119 root 10 -10 241.2g 1.3g 33016 S 0.0 0.3 0:00.00
> > dpdk_watchdog1
> > 13120 root 10 -10 241.2g 1.3g 33016 S 0.0 0.3 4:31.23
> > urcu2
> > 13127 root 10 -10 241.2g 1.3g 33016 S 0.0 0.3 0:00.04
> > ct_clean3
> > 13128 root 10 -10 241.2g 1.3g 33016 S 0.0 0.3 0:00.00
> > ipf_clean4
> > 13141 root 10 -10 241.2g 1.3g 33016 R 0.0 0.3 124:05.98
> > pmd-c20/id:13
> > 13142 root 10 -10 241.2g 1.3g 33016 R 0.0 0.3 124:05.98
> > pmd-c18/id:14
> > 13143 root 10 -10 241.2g 1.3g 33016 R 0.0 0.3 124:04.25
> > pmd-c35/id:15
> > 13145 root 10 -10 241.2g 1.3g 33016 R 0.0 0.3 124:05.99
> > pmd-c21/id:17
> > 13146 root 10 -10 241.2g 1.3g 33016 R 0.0 0.3 124:04.41
> > pmd-c05/id:18
> > 13147 root 10 -10 241.2g 1.3g 33016 R 0.0 0.3 124:05.98
> > pmd-c19/id:19
> > 13149 root 10 -10 241.2g 1.3g 33016 R 0.0 0.3 124:04.24
> > pmd-c36/id:21
> > 13150 root 10 -10 241.2g 1.3g 33016 R 0.0 0.3 124:04.32
> > pmd-c02/id:22
> > 13152 root 10 -10 241.2g 1.3g 33016 R 0.0 0.3 124:04.44
> > pmd-c04/id:24
> > 13153 root 10 -10 241.2g 1.3g 33016 R 0.0 0.3 124:04.24
> > pmd-c34/id:25
> > 13154 root 10 -10 241.2g 1.3g 33016 R 0.0 0.3 124:04.46
> > pmd-c03/id:26
> > 13157 root 10 -10 241.2g 1.3g 33016 S 0.0 0.3 0:00.04
> > vhost_reconn
> > 13158 root 10 -10 241.2g 1.3g 33016 S 0.0 0.3 0:00.69
> > vhost-events
> > 19115 root 10 -10 241.2g 1.3g 33016 S 0.0 0.3 0:00.00
> > handler54
> > 19116 root 10 -10 241.2g 1.3g 33016 S 0.0 0.3 0:00.00
> > handler55
> > 19117 root 10 -10 241.2g 1.3g 33016 S 0.0 0.3 0:00.00
> > handler56
> > 19118 root 10 -10 241.2g 1.3g 33016 S 0.0 0.3 0:00.00
> > handler61
> > 19119 root 10 -10 241.2g 1.3g 33016 S 0.0 0.3 0:00.00
> > handler57
> > 19120 root 10 -10 241.2g 1.3g 33016 S 0.0 0.3 1:49.58
> > revalidator58
> > 19121 root 10 -10 241.2g 1.3g 33016 S 0.0 0.3 1:43.02
> > revalidator59
> > 19122 root 10 -10 241.2g 1.3g 33016 S 0.0 0.3 1:40.94
> > revalidator60
> > _______________________________________________
> > discuss mailing list
> > [email protected]
> > https://mail.openvswitch.org/mailman/listinfo/ovs-discuss
> >
>
_______________________________________________
discuss mailing list
[email protected]
https://mail.openvswitch.org/mailman/listinfo/ovs-discuss