Allan, I increased the window size from 1 MB to 2 MB via
# ndd -set /dev/tcp tcp_cwnd_max 2097152 as suggested for 10GbE on the page mentioned by Nick, <http://www.solarisinternals.com/wiki/index.php/Networks>, but while that (or one of the other suggestions) improved aggregate throughput when running multiple streams, it didn't do much for the 100 MByte/s barrier I'm seeing for a single stream over 10GbE. Thanks, Chris Am 20.07.2014 um 15:34 schrieb Allan McAleavy via smartos-discuss <[email protected]>: > What's your tcp Window size I have seen in the past when using say veritas > vvr increasing this helps just a passing thought > > > On 20 Jul 2014 14:19, "Chris Ferebee via smartos-discuss" > <[email protected]> wrote: > > > > Nick, > > > > running 12 streams in parallel, I do see one CPU maxed out (core 0 as shown > > below). If I understand things correctly, it might be worthwhile to bump up > > rx_queue_number and tx_queue_number based on that. > > > > OTOH, when running just one stream, the one busy core doesn't seem to drop > > below 30% idle. I will still try increasing the number of queues next time > > I can schedule a reboot. > > > > Thanks, > > Chris > > > > ===== cpu stats while running 12 netcat streams: > > > > [root@90-e2-ba-00-2a-e2 ~]# mpstat 10 40 > > CPU minf mjf xcal intr ithr csw icsw migr smtx srw syscl usr sys wt idl > > 0 0 0 0 11671 9657 342 92 0 4263 0 0 0 100 0 0 > > 1 294 0 4 199 34 3876 9 73 771 0 3 0 5 0 95 > > 2 242 0 0 83 10 2384 0 26 576 0 2 0 2 0 98 > > 3 820 0 0 292 248 7959 1 18 1972 0 6 0 7 0 93 > > 4 267 0 0 24 3 2672 0 7 691 0 115 0 2 0 98 > > 5 983 0 0 36 2 8969 1 9 1857 0 114 0 5 0 95 > > 6 279 0 0 22 1 3800 13 17 2907 0 89029 3 14 0 82 > > 7 548 0 0 31 1 7160 21 30 3124 0 84476 3 17 0 80 > > 8 324 0 0 22 2 4290 15 20 3146 0 86657 3 14 0 83 > > 9 476 0 0 30 2 5989 18 25 3520 0 102163 4 20 0 > > 75 > > 10 261 0 0 24 2 3570 14 18 2579 0 88500 4 14 0 82 > > 11 480 0 0 42 4 6199 26 26 2838 0 81498 3 14 0 83 > > 12 2 0 0 18113 18090 41 0 8 432 0 0 0 10 0 > > 90 > > 13 137 0 0 5 0 1258 0 3 322 0 1 0 1 0 99 > > 14 706 0 0 43 28 7634 0 6 1743 0 44 0 5 0 95 > > 15 120 0 0 20 2 1441 0 3 328 0 3 0 1 0 99 > > 16 933 0 0 36 8 11714 0 9 2721 0 10 0 5 0 > > 95 > > 17 620 0 0 10049 10009 6046 0 12 1688 0 4 0 7 0 > > 93 > > 18 656 0 0 621 579 7786 29 31 2128 0 61547 3 13 0 85 > > 19 433 0 0 10657 10626 5522 14 26 1595 0 47377 2 19 0 > > 80 > > 20 706 0 0 133 93 8344 33 29 2536 0 60154 2 12 0 85 > > 21 486 0 0 15933 15904 5968 10 28 1611 0 38225 2 25 0 > > 74 > > 22 409 0 0 613 576 4730 26 28 3327 0 97758 4 17 0 79 > > 23 179 0 0 22 2 2528 15 18 2827 0 96562 4 15 0 81 > > > > > > Am 19.07.2014 um 23:16 schrieb Nick Perry via smartos-discuss > > <[email protected]>: > > > > > Hi Chris. > > > > > > How much improvement do you get with jumbo frames? > > > > > > Can you achieve significantly higher output if you try multiple streams > > > in parallel? > > > > > > During the test are there any CPU cores with very low idle time? > > > > > > Depending on the answers to the above it might be interesting to see if > > > there is any improvement by increasing rx_queue_number and > > > tx_queue_number on the ixgbe driver. > > > > > > Regards, > > > > > > Nick > > > > > > > > > On 19 July 2014 14:42, Chris Ferebee via smartos-discuss > > > <[email protected]> wrote: > > > > > > I'm trying to debug a network performance issue. > > > > > > I have two servers running SmartOS (20140613T024634Z and > > > 20140501T225642Z), one is a Supermicro dual Xeon E5649 (64 GB RAM) and > > > the other is a dual Xeon E5-2620v2 (128 GB RAM). Each has an Intel > > > X520-DA1 10GbE card, and they are both connected to 10GbE ports on a > > > NetGear GS752TXS switch. > > > > > > The switch reports 10GbE links: > > > > > > 1/xg49 Enable 10G Full 10G Full Link Up > > > Enable 1518 20:0C:C8:46:C8:3E 49 49 > > > 1/xg50 Enable 10G Full 10G Full Link Up > > > Enable 1518 20:0C:C8:46:C8:3E 50 50 > > > > > > as do both hosts: > > > > > > [root@90-e2-ba-00-2a-e2 ~]# dladm show-phys > > > LINK MEDIA STATE SPEED DUPLEX DEVICE > > > igb0 Ethernet down 0 > > > half igb0 > > > igb1 Ethernet down 0 > > > half igb1 > > > ixgbe0 Ethernet up 10000 full > > > ixgbe0 > > > > > > [root@00-1b-21-bf-e1-b4 ~]# dladm show-phys > > > LINK MEDIA STATE SPEED DUPLEX DEVICE > > > igb0 Ethernet down 0 > > > half igb0 > > > ixgbe0 Ethernet up 10000 full > > > ixgbe0 > > > igb1 Ethernet down 0 > > > half igb1 > > > > > > Per dladm show-linkprop, maxbw is not set on either of the net0 vnic > > > interfaces. > > > > > > And yet, as measured via netcat, throughput is just below 1 Gbit/s: > > > > > > [root@90-e2-ba-00-2a-e2 ~]# time cat /zones/test/10gb | nc -v -v -n > > > 192.168.168.5 8888 > > > Connection to 192.168.168.5 8888 port [tcp/*] succeeded! > > > > > > real 1m34.662s > > > user 0m11.422s > > > sys 1m53.957s > > > > > > (In this test, 10gb is a test file that is warm in RAM and transfers via > > > dd to /dev/null at approx. 2.4 GByte/s.) > > > > > > What could be causing the slowdown, and how might I go about debugging > > > this? > > > > > > FTR, disk throughput, while not an issue here, appears to be perfectly > > > reasonable, approx. 900 MB/s read performance. > > > > > > Thanks for any pointers! > > > > > > Chris > > > > > > > > > > > > > > > ------------------------------------------- > > > smartos-discuss > > > Archives: https://www.listbox.com/member/archive/184463/=now > > > RSS Feed: > > > https://www.listbox.com/member/archive/rss/184463/22416839-083bd2e9 > > > Modify Your Subscription: https://www.listbox.com/member/?& > > > Powered by Listbox: http://www.listbox.com > > > > > > smartos-discuss | Archives | Modify Your Subscription > > > > > > > > ------------------------------------------- > > smartos-discuss > > Archives: https://www.listbox.com/member/archive/184463/=now > > RSS Feed: > > https://www.listbox.com/member/archive/rss/184463/23685903-e846680a > > Modify Your Subscription: https://www.listbox.com/member/?& > > Powered by Listbox: http://www.listbox.com > > smartos-discuss | Archives | Modify Your Subscription ------------------------------------------- smartos-discuss Archives: https://www.listbox.com/member/archive/184463/=now RSS Feed: https://www.listbox.com/member/archive/rss/184463/25769125-55cfbc00 Modify Your Subscription: https://www.listbox.com/member/?member_id=25769125&id_secret=25769125-7688e9fb Powered by Listbox: http://www.listbox.com
