Hello Alex, Thanks for your immediate reply.
Alex, Can you please test the same performance with the same flow by enabling 2 cores of the server only.And share the performance report with me of OVS VS Kernel Bridge. On Wed, Dec 10, 2014 at 1:10 PM, Alex Wang <al...@nicira.com> wrote: > Hey Rahul, > > The warning means the handler threads was busying running for 16 seconds > before calling time_poll() again (in other words, before trying to sleep > again). > > The log showed suspiciously long system running. Definitely not expected, > > Could you confirm the ovs kernel module version via 'modinfo openvswitch'? > Never seen such warning with 2.3 kernel module, > > Thanks, > Alex Wang, > > On Tue, Dec 9, 2014 at 10:47 PM, Rahul Arora <rahul1991.ar...@gmail.com> > wrote: > >> Alex, >> >> I am getting these logs when i am getting 100% CPU usage. >> >> Please have a look on the RED part. >> >> ====================================== >> 2014-11-25T06:04:37.604Z|00095|timeval(handler5)|WARN|Unreasonably long >> 16127ms poll interval (260ms user, 17320ms system) >> 2014-11-25T06:04:37.604Z|00096|timeval(handler5)|WARN|faults: 8062 minor, >> 0 major >> 2014-11-25T06:04:37.604Z|00097|timeval(handler5)|WARN|context switches: >> 2424 voluntary, 670 involuntary >> 2014-11-25T06:04:37.604Z|00098|coverage(handler5)|INFO|Event coverage, >> avg rate over last: 5 seconds, last minute, last hour, hash=9b29cbda: >> 2014-11-25T06:04:37.604Z|00099|coverage(handler5)|INFO|netlink_sent >> 211.0/sec 77.300/sec 3.8958/sec total: 15051 >> 2014-11-25T06:04:37.604Z|00100|coverage(handler5)|INFO|netlink_recv_jumbo >> 0.4/sec 0.400/sec 0.0536/sec total: 195 >> 2014-11-25T06:04:37.604Z|00101|coverage(handler5)|INFO|netlink_received >> 213.0/sec 78.983/sec 4.1219/sec total: 15873 >> 2014-11-25T06:04:37.604Z|00102|coverage(handler5)|INFO|netlink_overflow >> 0.0/sec 0.033/sec 0.0011/sec total: 4 >> 2014-11-25T06:04:37.604Z|00103|coverage(handler5)|INFO|netdev_set_ethtool >> 0.0/sec 0.000/sec 0.0006/sec total: 2 >> 2014-11-25T06:04:37.604Z|00104|coverage(handler5)|INFO|netdev_get_ethtool >> 0.0/sec 0.000/sec 0.0028/sec total: 10 >> 2014-11-25T06:04:37.605Z|00105|coverage(handler5)|INFO|netdev_set_hwaddr >> 0.0/sec 0.000/sec 0.0008/sec total: 3 >> 2014-11-25T06:04:37.605Z|00106|coverage(handler5)|INFO|netdev_get_hwaddr >> 0.0/sec 0.000/sec 0.0008/sec total: 3 >> 2014-11-25T06:04:37.605Z|00107|coverage(handler5)|INFO|netdev_get_ifindex >> 0.0/sec 0.000/sec 0.0008/sec total: 3 >> 2014-11-25T06:04:37.605Z|00108|coverage(handler5)|INFO|netdev_set_policing >> 0.0/sec 0.000/sec 0.0017/sec total: 6 >> 2014-11-25T06:04:37.605Z|00109|coverage(handler5)|INFO|vconn_sent >> 0.0/sec 0.000/sec 0.0014/sec total: 5 >> 2014-11-25T06:04:37.605Z|00110|coverage(handler5)|INFO|vconn_received >> 0.0/sec 0.000/sec 0.0019/sec total: 7 >> 2014-11-25T06:04:37.605Z|00111|coverage(handler5)|INFO|util_xalloc >> 396.6/sec 367.317/sec 50.7611/sec total: 184603 >> 2014-11-25T06:04:37.605Z|00112|coverage(handler5)|INFO|stream_open >> 0.0/sec 0.000/sec 0.0003/sec total: 1 >> 2014-11-25T06:04:37.605Z|00113|coverage(handler5)|INFO|pstream_open >> 0.0/sec 0.000/sec 0.0008/sec total: 3 >> 2014-11-25T06:04:37.605Z|00114|coverage(handler5)|INFO|rconn_sent >> 0.0/sec 0.000/sec 0.0008/sec total: 3 >> 2014-11-25T06:04:37.605Z|00115|coverage(handler5)|INFO|rconn_queued >> 0.0/sec 0.000/sec 0.0008/sec total: 3 >> 2014-11-25T06:04:37.605Z|00116|coverage(handler5)|INFO|poll_zero_timeout >> 0.0/sec 0.000/sec 0.0031/sec total: 11 >> 2014-11-25T06:04:37.605Z|00117|coverage(handler5)|INFO|poll_fd_wait >> 30.0/sec 29.617/sec 4.1908/sec total: 15201 >> 2014-11-25T06:04:37.605Z|00118|coverage(handler5)|INFO|txn_success >> 0.0/sec 0.000/sec 0.0017/sec total: 6 >> 2014-11-25T06:04:37.605Z|00119|coverage(handler5)|INFO|txn_incomplete >> 0.2/sec 0.183/sec 0.0278/sec total: 101 >> 2014-11-25T06:04:37.605Z|00120|coverage(handler5)|INFO|txn_unchanged >> 0.0/sec 0.017/sec 0.0075/sec total: 27 >> 2014-11-25T06:04:37.605Z|00121|coverage(handler5)|INFO|netdev_get_stats >> 0.6/sec 0.600/sec 0.0811/sec total: 295 >> 2014-11-25T06:04:37.605Z|00122|coverage(handler5)|INFO|mac_learning_expired >> 0.0/sec 0.000/sec 0.0003/sec total: 1 >> 2014-11-25T06:04:37.605Z|00123|coverage(handler5)|INFO|mac_learning_learned >> 0.0/sec 0.000/sec 0.0003/sec total: 1 >> 2014-11-25T06:04:37.605Z|00124|coverage(handler5)|INFO|hmap_expand >> 18.4/sec 17.883/sec 2.5731/sec total: 9337 >> 2014-11-25T06:04:37.605Z|00125|coverage(handler5)|INFO|hmap_pathological >> 0.0/sec 0.000/sec 0.0053/sec total: 19 >> 2014-11-25T06:04:37.605Z|00126|coverage(handler5)|INFO|flow_extract >> 200.0/sec 70.417/sec 3.1047/sec total: 12150 >> 2014-11-25T06:04:37.605Z|00127|coverage(handler5)|INFO|dpif_execute >> 0.0/sec 0.000/sec 0.0003/sec total: 1 >> 2014-11-25T06:04:37.605Z|00128|coverage(handler5)|INFO|dpif_flow_put >> 0.0/sec 0.000/sec 0.0006/sec total: 2 >> 2014-11-25T06:04:37.605Z|00129|coverage(handler5)|INFO|dpif_flow_flush >> 0.0/sec 0.000/sec 0.0006/sec total: 2 >> 2014-11-25T06:04:37.605Z|00130|coverage(handler5)|INFO|dpif_port_add >> 0.0/sec 0.000/sec 0.0008/sec total: 3 >> 2014-11-25T06:04:37.605Z|00131|coverage(handler5)|INFO|xlate_actions >> 4.0/sec 1.433/sec 0.0656/sec total: 256 >> 2014-11-25T06:04:37.605Z|00132|coverage(handler5)|INFO|rev_mac_learning >> 0.0/sec 0.000/sec 0.0006/sec total: 2 >> 2014-11-25T06:04:37.605Z|00133|coverage(handler5)|INFO|rev_flow_table >> 0.0/sec 0.000/sec 0.0006/sec total: 2 >> 2014-11-25T06:04:37.605Z|00134|coverage(handler5)|INFO|rev_port_toggled >> 0.0/sec 0.000/sec 0.0003/sec total: 1 >> 2014-11-25T06:04:37.605Z|00135|coverage(handler5)|INFO|rev_reconfigure >> 0.0/sec 0.000/sec 0.0008/sec total: 3 >> 2014-11-25T06:04:37.605Z|00136|coverage(handler5)|INFO|ofproto_update_port >> 0.0/sec 0.000/sec 0.0025/sec total: 9 >> 2014-11-25T06:04:37.605Z|00137|coverage(handler5)|INFO|ofproto_recv_openflow >> 0.0/sec 0.000/sec 0.0014/sec total: 5 >> 2014-11-25T06:04:37.605Z|00138|coverage(handler5)|INFO|ofproto_flush >> 0.0/sec 0.000/sec 0.0003/sec total: 1 >> 2014-11-25T06:04:37.605Z|00139|coverage(handler5)|INFO|bridge_reconfigure >> 0.0/sec 0.000/sec 0.0011/sec total: 4 >> 2014-11-25T06:04:37.605Z|00140|coverage(handler5)|INFO|49 events never hit >> ======================================================================== >> >> >> Do you have some idea how i can optimize it? >> >> On Wed, Dec 10, 2014 at 11:41 AM, Alex Wang <al...@nicira.com> wrote: >> >>> >>> I believe it has 16 cores, with hyperthreading disabled, >>> >>> On Tue, Dec 9, 2014 at 9:37 PM, Rahul Arora <rahul1991.ar...@gmail.com> >>> wrote: >>> >>>> Hi Alex, >>>> >>>> The platform which you are using for your performance testing is of how >>>> many cores?? >>>> >>>> On Tue, Dec 9, 2014 at 10:45 PM, Alex Wang <al...@nicira.com> wrote: >>>> >>>>> Hey Rahul, >>>>> >>>>> The kernel version should not make a difference, >>>>> >>>>> I'm curious, since ovs-2.3.0 is multithreaded, not sure how do you >>>>> measure >>>>> the cpu usage to be 100%? >>>>> >>>>> Also, could you post the full `mpstat -P ALL` output, so we can see the >>>>> distribution of cpu usage, >>>>> >>>>> >>>>> Thanks, >>>>> Alex Wang, >>>>> >>>>> >>>>> >>>>> On Tue, Dec 9, 2014 at 1:49 AM, Rahul Arora <rahul1991.ar...@gmail.com >>>>> > wrote: >>>>> >>>>>> Hi Alex, >>>>>> >>>>>> Thanks for your help. >>>>>> >>>>>> I am using the following command to test CPU usage. >>>>>> >>>>>> mpstat -P ALL 1 >>>>>> >>>>>> I was using 3.12 kernel version and you are using 3.13.0-30-generic >>>>>> version.Please let me know if this can be the issue?? >>>>>> >>>>>> >>>>>> >>>>>> On Sat, Dec 6, 2014 at 5:04 AM, Alex Wang <al...@nicira.com> wrote: >>>>>> >>>>>>> Hey Rahul, >>>>>>> >>>>>>> Could you help confirm the following: >>>>>>> >>>>>>> 1. did you use vm or bare metal machine for testing? >>>>>>> >>>>>>> 2. how do you measure the cpu utilization of ovs and kernel bridge? >>>>>>> >>>>>>> >>>>>>> We ran some throughput tests, and we observed constant better >>>>>>> performance over the Linux bridge. >>>>>>> >>>>>>> Below is a set of our own throughput test result using ovs >>>>>>> branch-2.3. >>>>>>> >>>>>>> Plateform: >>>>>>> Server Dell R620, running Ubuntu 14.04 (3.13.0-30-generic), NIC is >>>>>>> Intel X540. >>>>>>> >>>>>>> Flow: >>>>>>> Single flow, "in_port=1,actions:output=2" >>>>>>> >>>>>>> Traffic type: >>>>>>> - unidirectional TCP throughput using Spirent (1 flow and 128 flows). >>>>>>> >>>>>>> >>>>>>> *Throughput view (unit: Gbit/s):* >>>>>>> >>>>>>> Linux Bridge, >>>>>>> Frame Size1 flow128 flow780.6021.5121280.9812.1892561.8155.433512 >>>>>>> 3.2525.29910246.65810.015149.6849.98790009.92110.0 >>>>>>> >>>>>>> OVS Branch-2.3: >>>>>>> Frame Size1 flow128 flow780.78.2581281.09810.02561.89810.05123.694 >>>>>>> 10.010247.1910.015149.8849.98790009.89910.0 >>>>>>> >>>>>>> >>>>>>> >>>>>>> *Frame Per Second view (unit: Frames/s):* >>>>>>> >>>>>>> Linux Bridge, >>>>>>> Frame Size1 flow128 flow7876805719290121288289121849112256822070 >>>>>>> 246063051276405812450201024797193119731815147891418138029000137491 >>>>>>> 138581 >>>>>>> >>>>>>> OVS Branch-2.3: >>>>>>> Frame Size1 flow128 flow78892857105337071289272998445946256859386 >>>>>>> 452898551286805523496241024860881119731815148054128138029000137181 >>>>>>> 138581 >>>>>>> >>>>>>> >>>>>>> Thanks, >>>>>>> Alex Wang, >>>>>>> >>>>>>> On Wed, Nov 26, 2014 at 10:35 PM, Rahul Arora < >>>>>>> rahul1991.ar...@gmail.com> wrote: >>>>>>> >>>>>>>> Hi Team, >>>>>>>> >>>>>>>> We are doing comparison of throughput and CPU consumption between >>>>>>>> OVS 2.3.0 and kernel bridge with different packet size. >>>>>>>> >>>>>>>> We are observing huge difference in performance. With frame size 64 >>>>>>>> and 128 bytes unidirectional traffic from port1 to port2 below are the >>>>>>>> numbers. >>>>>>>> >>>>>>>> >>>>>>>> *OVS 2.3.0 (Dual core with matching flow in kernel space, kernel >>>>>>>> 3.12)* >>>>>>>> >>>>>>>> * kernel bridge(Dual Core System, kernel 3.12) * *FRAME SIZE* >>>>>>>> *Throughput >>>>>>>> unidirectional (Mbps)* *CPU Usage %* *VS* *FRAME SIZE* *Throughput >>>>>>>> (Mbps)* *CPU Usage %* 64 375 100 64 487 100 128 747 100 128 864 >>>>>>>> 40 256 927 10 256 927 5 320 941 8 320 941 4 384 950 6 384 950 4 >>>>>>>> 448 957 4 448 957 3 512 962 3 512 962 3 1024 980 1 1024 980 1 >>>>>>>> 1500 986 1 1500 986 1 >>>>>>>> We have matching flow in kernel space with in_port=1 and >>>>>>>> action=output:2 and flow is matching in kernel space >>>>>>>> >>>>>>>> How we can improve the performance of OVS i.e. increasing the >>>>>>>> throughput and decrease CPU consumption with lower frame size. >>>>>>>> >>>>>>>> _______________________________________________ >>>>>>>> discuss mailing list >>>>>>>> disc...@openvswitch.org >>>>>>>> http://openvswitch.org/mailman/listinfo/discuss >>>>>>>> >>>>>>>> >>>>>>> >>>>>> >>>>> >>>> >>> >> >
_______________________________________________ openflow-discuss mailing list openflow-discuss@lists.stanford.edu https://mailman.stanford.edu/mailman/listinfo/openflow-discuss