Would you report your speeds simply changing the protocol between tcp and
udp?


On Sun, Jul 20, 2014 at 8:08 AM, Chris Ferebee via smartos-discuss <
[email protected]> wrote:

> Allan,
>
> I increased the window size from 1 MB to 2 MB via
>
> # ndd -set /dev/tcp tcp_cwnd_max 2097152
>
> as suggested for 10GbE on the page mentioned by Nick, <
> http://www.solarisinternals.com/wiki/index.php/Networks>, but while that
> (or one of the other suggestions) improved aggregate throughput when
> running multiple streams, it didn't do much for the 100 MByte/s barrier I'm
> seeing for a single stream over 10GbE.
>
> Thanks,
> Chris
>
>
>
> Am 20.07.2014 um 15:34 schrieb Allan McAleavy via smartos-discuss <
> [email protected]>:
>
> > What's your tcp Window size I have seen in the past when using say
> veritas vvr increasing this helps just a passing thought
> >
> >
> > On 20 Jul 2014 14:19, "Chris Ferebee via smartos-discuss" <
> [email protected]> wrote:
> > >
> > > Nick,
> > >
> > > running 12 streams in parallel, I do see one CPU maxed out (core 0 as
> shown below). If I understand things correctly, it might be worthwhile to
> bump up rx_queue_number and tx_queue_number based on that.
> > >
> > > OTOH, when running just one stream, the one busy core doesn't seem to
> drop below 30% idle. I will still try increasing the number of queues next
> time I can schedule a reboot.
> > >
> > > Thanks,
> > > Chris
> > >
> > > ===== cpu stats while running 12 netcat streams:
> > >
> > > [root@90-e2-ba-00-2a-e2 ~]# mpstat 10 40
> > > CPU minf mjf xcal  intr ithr  csw icsw migr smtx  srw syscl  usr sys
>  wt idl
> > >   0    0   0    0 11671 9657  342   92    0 4263    0     0    0 100
> 0   0
> > >   1  294   0    4   199   34 3876    9   73  771    0     3    0   5
> 0  95
> > >   2  242   0    0    83   10 2384    0   26  576    0     2    0   2
> 0  98
> > >   3  820   0    0   292  248 7959    1   18 1972    0     6    0   7
> 0  93
> > >   4  267   0    0    24    3 2672    0    7  691    0   115    0   2
> 0  98
> > >   5  983   0    0    36    2 8969    1    9 1857    0   114    0   5
> 0  95
> > >   6  279   0    0    22    1 3800   13   17 2907    0 89029    3  14
> 0  82
> > >   7  548   0    0    31    1 7160   21   30 3124    0 84476    3  17
> 0  80
> > >   8  324   0    0    22    2 4290   15   20 3146    0 86657    3  14
> 0  83
> > >   9  476   0    0    30    2 5989   18   25 3520    0 102163    4  20
>   0  75
> > >  10  261   0    0    24    2 3570   14   18 2579    0 88500    4  14
> 0  82
> > >  11  480   0    0    42    4 6199   26   26 2838    0 81498    3  14
> 0  83
> > >  12    2   0    0 18113 18090   41    0    8  432    0     0    0  10
>   0  90
> > >  13  137   0    0     5    0 1258    0    3  322    0     1    0   1
> 0  99
> > >  14  706   0    0    43   28 7634    0    6 1743    0    44    0   5
> 0  95
> > >  15  120   0    0    20    2 1441    0    3  328    0     3    0   1
> 0  99
> > >  16  933   0    0    36    8 11714    0    9 2721    0    10    0   5
>   0  95
> > >  17  620   0    0 10049 10009 6046    0   12 1688    0     4    0   7
>   0  93
> > >  18  656   0    0   621  579 7786   29   31 2128    0 61547    3  13
> 0  85
> > >  19  433   0    0 10657 10626 5522   14   26 1595    0 47377    2  19
>   0  80
> > >  20  706   0    0   133   93 8344   33   29 2536    0 60154    2  12
> 0  85
> > >  21  486   0    0 15933 15904 5968   10   28 1611    0 38225    2  25
>   0  74
> > >  22  409   0    0   613  576 4730   26   28 3327    0 97758    4  17
> 0  79
> > >  23  179   0    0    22    2 2528   15   18 2827    0 96562    4  15
> 0  81
> > >
> > >
> > > Am 19.07.2014 um 23:16 schrieb Nick Perry via smartos-discuss <
> [email protected]>:
> > >
> > > > Hi Chris.
> > > >
> > > > How much improvement do you get with jumbo frames?
> > > >
> > > > Can you achieve significantly higher output if you try multiple
> streams in parallel?
> > > >
> > > > During the test are there any CPU cores with very low idle time?
> > > >
> > > > Depending on the answers to the above it might be interesting to see
> if there is any improvement by increasing rx_queue_number and
> tx_queue_number on the ixgbe driver.
> > > >
> > > > Regards,
> > > >
> > > > Nick
> > > >
> > > >
> > > > On 19 July 2014 14:42, Chris Ferebee via smartos-discuss <
> [email protected]> wrote:
> > > >
> > > > I'm trying to debug a network performance issue.
> > > >
> > > > I have two servers running SmartOS (20140613T024634Z and
> 20140501T225642Z), one is a Supermicro dual Xeon E5649 (64 GB RAM) and the
> other is a dual Xeon E5-2620v2 (128 GB RAM). Each has an Intel X520-DA1
> 10GbE card, and they are both connected to 10GbE ports on a NetGear
> GS752TXS switch.
> > > >
> > > > The switch reports 10GbE links:
> > > >
> > > > 1/xg49                  Enable  10G Full        10G Full        Link
> Up Enable  1518    20:0C:C8:46:C8:3E       49      49
> > > > 1/xg50                  Enable  10G Full        10G Full        Link
> Up Enable  1518    20:0C:C8:46:C8:3E       50      50
> > > >
> > > > as do both hosts:
> > > >
> > > > [root@90-e2-ba-00-2a-e2 ~]# dladm show-phys
> > > > LINK    MEDIA           STATE   SPEED           DUPLEX
>  DEVICE
> > > > igb0            Ethernet                down    0
>     half                    igb0
> > > > igb1            Ethernet                down    0
>     half                    igb1
> > > > ixgbe0  Ethernet                up              10000           full
>                    ixgbe0
> > > >
> > > > [root@00-1b-21-bf-e1-b4 ~]# dladm show-phys
> > > > LINK    MEDIA           STATE   SPEED           DUPLEX
>  DEVICE
> > > > igb0            Ethernet                down    0
>     half                    igb0
> > > > ixgbe0  Ethernet                up              10000           full
>                    ixgbe0
> > > > igb1            Ethernet                down    0
>     half                    igb1
> > > >
> > > > Per dladm show-linkprop, maxbw is not set on either of the net0 vnic
> interfaces.
> > > >
> > > > And yet, as measured via netcat, throughput is just below 1 Gbit/s:
> > > >
> > > > [root@90-e2-ba-00-2a-e2 ~]# time cat /zones/test/10gb | nc -v -v -n
> 192.168.168.5 8888
> > > > Connection to 192.168.168.5 8888 port [tcp/*] succeeded!
> > > >
> > > > real            1m34.662s
> > > > user            0m11.422s
> > > > sys             1m53.957s
> > > >
> > > > (In this test, 10gb is a test file that is warm in RAM and transfers
> via dd to /dev/null at approx. 2.4 GByte/s.)
> > > >
> > > > What could be causing the slowdown, and how might I go about
> debugging this?
> > > >
> > > > FTR, disk throughput, while not an issue here, appears to be
> perfectly reasonable, approx. 900 MB/s read performance.
> > > >
> > > > Thanks for any pointers!
> > > >
> > > > Chris
> > > >
> > > >
> > > >
> > > >
> > > > -------------------------------------------
> > > > smartos-discuss
> > > > Archives: https://www.listbox.com/member/archive/184463/=now
> > > > RSS Feed:
> https://www.listbox.com/member/archive/rss/184463/22416839-083bd2e9
> > > > Modify Your Subscription: https://www.listbox.com/member/?&;
> > > > Powered by Listbox: http://www.listbox.com
> > > >
> > > > smartos-discuss | Archives  | Modify Your Subscription
> > >
> > >
> > >
> > > -------------------------------------------
> > > smartos-discuss
> > > Archives: https://www.listbox.com/member/archive/184463/=now
> > > RSS Feed:
> https://www.listbox.com/member/archive/rss/184463/23685903-e846680a
> > > Modify Your Subscription: https://www.listbox.com/member/?&;
> > > Powered by Listbox: http://www.listbox.com
> >
> > smartos-discuss | Archives  | Modify Your Subscription
>
>
>
> -------------------------------------------
> smartos-discuss
> Archives: https://www.listbox.com/member/archive/184463/=now
> RSS Feed:
> https://www.listbox.com/member/archive/rss/184463/24535378-4ca6b74b
> Modify Your Subscription:
> https://www.listbox.com/member/?&;
> Powered by Listbox: http://www.listbox.com
>



-------------------------------------------
smartos-discuss
Archives: https://www.listbox.com/member/archive/184463/=now
RSS Feed: https://www.listbox.com/member/archive/rss/184463/25769125-55cfbc00
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=25769125&id_secret=25769125-7688e9fb
Powered by Listbox: http://www.listbox.com

Reply via email to