Nick,

running 12 streams in parallel, I do see one CPU maxed out (core 0 as shown 
below). If I understand things correctly, it might be worthwhile to bump up 
rx_queue_number and tx_queue_number based on that.

OTOH, when running just one stream, the one busy core doesn't seem to drop 
below 30% idle. I will still try increasing the number of queues next time I 
can schedule a reboot.

Thanks,
Chris

===== cpu stats while running 12 netcat streams:

[root@90-e2-ba-00-2a-e2 ~]# mpstat 10 40
CPU minf mjf xcal  intr ithr  csw icsw migr smtx  srw syscl  usr sys  wt idl
  0    0   0    0 11671 9657  342   92    0 4263    0     0    0 100   0   0
  1  294   0    4   199   34 3876    9   73  771    0     3    0   5   0  95
  2  242   0    0    83   10 2384    0   26  576    0     2    0   2   0  98
  3  820   0    0   292  248 7959    1   18 1972    0     6    0   7   0  93
  4  267   0    0    24    3 2672    0    7  691    0   115    0   2   0  98
  5  983   0    0    36    2 8969    1    9 1857    0   114    0   5   0  95
  6  279   0    0    22    1 3800   13   17 2907    0 89029    3  14   0  82
  7  548   0    0    31    1 7160   21   30 3124    0 84476    3  17   0  80
  8  324   0    0    22    2 4290   15   20 3146    0 86657    3  14   0  83
  9  476   0    0    30    2 5989   18   25 3520    0 102163    4  20   0  75
 10  261   0    0    24    2 3570   14   18 2579    0 88500    4  14   0  82
 11  480   0    0    42    4 6199   26   26 2838    0 81498    3  14   0  83
 12    2   0    0 18113 18090   41    0    8  432    0     0    0  10   0  90
 13  137   0    0     5    0 1258    0    3  322    0     1    0   1   0  99
 14  706   0    0    43   28 7634    0    6 1743    0    44    0   5   0  95
 15  120   0    0    20    2 1441    0    3  328    0     3    0   1   0  99
 16  933   0    0    36    8 11714    0    9 2721    0    10    0   5   0  95
 17  620   0    0 10049 10009 6046    0   12 1688    0     4    0   7   0  93
 18  656   0    0   621  579 7786   29   31 2128    0 61547    3  13   0  85
 19  433   0    0 10657 10626 5522   14   26 1595    0 47377    2  19   0  80
 20  706   0    0   133   93 8344   33   29 2536    0 60154    2  12   0  85
 21  486   0    0 15933 15904 5968   10   28 1611    0 38225    2  25   0  74
 22  409   0    0   613  576 4730   26   28 3327    0 97758    4  17   0  79
 23  179   0    0    22    2 2528   15   18 2827    0 96562    4  15   0  81


Am 19.07.2014 um 23:16 schrieb Nick Perry via smartos-discuss 
<[email protected]>:

> Hi Chris.
> 
> How much improvement do you get with jumbo frames?
> 
> Can you achieve significantly higher output if you try multiple streams in 
> parallel?
> 
> During the test are there any CPU cores with very low idle time?
> 
> Depending on the answers to the above it might be interesting to see if there 
> is any improvement by increasing rx_queue_number and tx_queue_number on the 
> ixgbe driver.
> 
> Regards,
> 
> Nick
> 
> 
> On 19 July 2014 14:42, Chris Ferebee via smartos-discuss 
> <[email protected]> wrote:
> 
> I'm trying to debug a network performance issue.
> 
> I have two servers running SmartOS (20140613T024634Z and 20140501T225642Z), 
> one is a Supermicro dual Xeon E5649 (64 GB RAM) and the other is a dual Xeon 
> E5-2620v2 (128 GB RAM). Each has an Intel X520-DA1 10GbE card, and they are 
> both connected to 10GbE ports on a NetGear GS752TXS switch.
> 
> The switch reports 10GbE links:
> 
> 1/xg49                  Enable  10G Full        10G Full        Link Up 
> Enable  1518    20:0C:C8:46:C8:3E       49      49
> 1/xg50                  Enable  10G Full        10G Full        Link Up 
> Enable  1518    20:0C:C8:46:C8:3E       50      50
> 
> as do both hosts:
> 
> [root@90-e2-ba-00-2a-e2 ~]# dladm show-phys
> LINK    MEDIA           STATE   SPEED           DUPLEX          DEVICE
> igb0            Ethernet                down    0                       half  
>                   igb0
> igb1            Ethernet                down    0                       half  
>                   igb1
> ixgbe0  Ethernet                up              10000           full          
>           ixgbe0
> 
> [root@00-1b-21-bf-e1-b4 ~]# dladm show-phys
> LINK    MEDIA           STATE   SPEED           DUPLEX          DEVICE
> igb0            Ethernet                down    0                       half  
>                   igb0
> ixgbe0  Ethernet                up              10000           full          
>           ixgbe0
> igb1            Ethernet                down    0                       half  
>                   igb1
> 
> Per dladm show-linkprop, maxbw is not set on either of the net0 vnic 
> interfaces.
> 
> And yet, as measured via netcat, throughput is just below 1 Gbit/s:
> 
> [root@90-e2-ba-00-2a-e2 ~]# time cat /zones/test/10gb | nc -v -v -n 
> 192.168.168.5 8888
> Connection to 192.168.168.5 8888 port [tcp/*] succeeded!
> 
> real            1m34.662s
> user            0m11.422s
> sys             1m53.957s
> 
> (In this test, 10gb is a test file that is warm in RAM and transfers via dd 
> to /dev/null at approx. 2.4 GByte/s.)
> 
> What could be causing the slowdown, and how might I go about debugging this?
> 
> FTR, disk throughput, while not an issue here, appears to be perfectly 
> reasonable, approx. 900 MB/s read performance.
> 
> Thanks for any pointers!
> 
> Chris
> 
> 
> 
> 
> -------------------------------------------
> smartos-discuss
> Archives: https://www.listbox.com/member/archive/184463/=now
> RSS Feed: https://www.listbox.com/member/archive/rss/184463/22416839-083bd2e9
> Modify Your Subscription: https://www.listbox.com/member/?&;
> Powered by Listbox: http://www.listbox.com
> 
> smartos-discuss | Archives  | Modify Your Subscription         



-------------------------------------------
smartos-discuss
Archives: https://www.listbox.com/member/archive/184463/=now
RSS Feed: https://www.listbox.com/member/archive/rss/184463/25769125-55cfbc00
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=25769125&id_secret=25769125-7688e9fb
Powered by Listbox: http://www.listbox.com

Reply via email to