Re: from the academic side of the house

2007-04-29 Thread JP Velders

On Tue, 24 Apr 2007 [EMAIL PROTECTED] wrote:

 Date: Tue, 24 Apr 2007 15:36:51 +
 From: [EMAIL PROTECTED]
 Subject: from the academic side of the house

 For the first set of IPv6 records, a team from the University of Tokyo, WIDE
 Project, NTT Communications, JGN2, SURFnet, CANARIE, Pacific Northwest
 Gigapop and other institutions collaborated to create a network path over
 30,000 kilometers in distance, crossing 6 international networks - over 3/4
 the circumference of the Earth. In doing so, the team successfully
 transferred data in the single and multi-stream categories at a rate of 7.67
 Gbps which is equal to 230,100 terabit-meters per second (Tb-m/s).  This
 record setting attempt leveraged standard TCP to achieve the new mark.

Mind you, those crazy Japanese do this every year between christmas 
and newyear... ;) Most of the pipes they used also carry other 
research traffic throughout most of the year... This year was even 
more cumbersome because of some issues with the OC192's between 
Amsterdam and the USA...

Kind regards,
JP Velders


Re: from the academic side of the house

2007-04-29 Thread JP Velders


On Tue, 24 Apr 2007, Jim Shankland wrote:

 Date: Tue, 24 Apr 2007 09:24:13 -0700
 From: Jim Shankland [EMAIL PROTECTED]
 Subject: Re: from the academic side of the house

 (1) Do the throughput figures count only the data payload (i.e.,
 anything above the TCP layer), or all the bits from the protocol
 stack?  If the latter, it seems a little unreasonable to credit
 IPv6 with its own extra overhead -- though I'll concede that with
 jumbo datagrams, that's not all that much.

Data payload is counted as bytes transmitted and received by iperf. So 
application layer all the way.

 (2) Getting this kind of throughput seems to depend on a fast
 physical layer, plus some link-layer help (jumbo packets), plus
 careful TCP tuning to deal with the large bandwidth-delay product.

That last part has been researched for quite some time already, though 
mainly with long transatlantic layer 2 (Ethernet) paths mainly.

 The IP layer sits between the second and third of those three items.
 Is there something about IPv6 vs. IPv4 that specifically improves
 perfomance on this kind of test?  If so, what is it?

Not that was specificly mentioned for this test I believe...

Kind regards,
JP Velders


Re: from the academic side of the house

2007-04-29 Thread JP Velders

On Thu, 26 Apr 2007, Simon Leinen wrote:

 Date: Thu, 26 Apr 2007 19:11:37 +0200
 From: Simon Leinen [EMAIL PROTECTED]
 Subject: Re: from the academic side of the house

 [ ... ]
 Another host issue would be interrupts and CPU load for checksum, but
 most modern 10GE (and also GigE!) adapters offload checksum
 segmentation and reassembly, as well as checksum computation and
 validation to the adapter if the OS/driver supports it.

Correct, though sometimes the performance can vary wildly depending on 
driver and firmware. I had to turn off all the checksumming with 
certain driver versions for the Myrinet 10GE cards, later drivers 
performed OK.

Also, having multi-cpu and multi-core machines helps enormously, 
binding a NIC/driver to a certain core, and having the receiving 
process on the neighbouring core can dramatically increase throughput.

But per adapter/driver differences in which offloading actually 
increases performance sometimes still is a black art, necessitating 
painstaking testing...

 The adapters used in this record (Chelsio S310E) contain a full TOE
 (TCP Offload Engine) that can run the entire TCP state machine on the
 adapter, although I'm more sure whether they made use of that.

I believe that they did not use it. In the past we also ran into 
problems and performance dips when using the accellerated Chelsio 
cards. Without TOE accelleration things worked much smoother.

Kei's team did modify iperf such that they used mmap's to gain more of 
a zero-copy behaviour, depending on drivers this can also give much 
better performance, akin to say Myricom's claims for Myrinet...

Kind regards,
JP Velders


Re: from the academic side of the house

2007-04-29 Thread bmanning

On Sun, Apr 29, 2007 at 01:57:26PM +0200, JP Velders wrote:
 On Tue, 24 Apr 2007 [EMAIL PROTECTED] wrote:
 
  Date: Tue, 24 Apr 2007 15:36:51 +
  From: [EMAIL PROTECTED]
  Subject: from the academic side of the house
 
  For the first set of IPv6 records, a team from the University of Tokyo, WIDE
  Project, NTT Communications, JGN2, SURFnet, CANARIE, Pacific Northwest
  Gigapop and other institutions collaborated to create a network path over
  30,000 kilometers in distance, crossing 6 international networks - over 3/4
  the circumference of the Earth. In doing so, the team successfully
  transferred data in the single and multi-stream categories at a rate of 7.67
  Gbps which is equal to 230,100 terabit-meters per second (Tb-m/s).  This
  record setting attempt leveraged standard TCP to achieve the new mark.
 
 Mind you, those crazy Japanese do this every year between christmas 
 and newyear... ;) Most of the pipes they used also carry other 
 research traffic throughout most of the year... This year was even 
 more cumbersome because of some issues with the OC192's between 
 Amsterdam and the USA...
 
 Kind regards,
 JP Velders

we -love- the crazy Japanese doing this kind of stuff.
the US folks seemed to have lost momentum in the past decade.
while the pipes do get re-purposed on a regular basis, they
do tend to shake out interoperable problems, as you note above.

me, i await the spiral loop that includes the southern 
hemisphere ...

--bill


Re: from the academic side of the house

2007-04-26 Thread Simon Leinen

Tony Li writes:
 On Apr 25, 2007, at 2:55 PM, Simon Leinen wrote:
 Routing table lookups(*) are what's most relevant here, [...]

 Actually, what's most relevant here is the ability to get end-hosts
 to run at rate.  Packet forwarding at line rate has been
 demonstrated for quite awhile now.

That's true (although Steve's question was about the routers).

The host bottleneck for raw 10Gb/s transfers used to be bus bandwidth.
The 10GE adapters in most older land-speed record entries used the
slower PCI-X, while this entry was done with PCI Express (x8) adapters.

Another host issue would be interrupts and CPU load for checksum, but
most modern 10GE (and also GigE!) adapters offload checksum
segmentation and reassembly, as well as checksum computation and
validation to the adapter if the OS/driver supports it.

The adapters used in this record (Chelsio S310E) contain a full TOE
(TCP Offload Engine) that can run the entire TCP state machine on the
adapter, although I'm more sure whether they made use of that.
Details on

http://data-reservoir.adm.s.u-tokyo.ac.jp/lsr-200612-02/
-- 
Simon.


Re: from the academic side of the house

2007-04-25 Thread Steven M. Bellovin

On Tue, 24 Apr 2007 09:24:13 -0700
Jim Shankland [EMAIL PROTECTED] wrote:

 (2) Getting this kind of throughput seems to depend on a fast
 physical layer, plus some link-layer help (jumbo packets), plus
 careful TCP tuning to deal with the large bandwidth-delay product.
 The IP layer sits between the second and third of those three items.
 Is there something about IPv6 vs. IPv4 that specifically improves
 perfomance on this kind of test?  If so, what is it?

I wonder if the router forward v6 as fast.


--Steve Bellovin, http://www.cs.columbia.edu/~smb


Re: from the academic side of the house

2007-04-25 Thread Simon Leinen

Steven M Bellovin writes:
 Jim Shankland [EMAIL PROTECTED] wrote:

 (2) Getting this kind of throughput seems to depend on a fast
 physical layer, plus some link-layer help (jumbo packets), plus
 careful TCP tuning to deal with the large bandwidth-delay product.
 The IP layer sits between the second and third of those three items.
 Is there something about IPv6 vs. IPv4 that specifically improves
 perfomance on this kind of test?  If so, what is it?

 I wonder if the router forward v6 as fast.

In the 10 Gb/s space (sufficient for these records, and I'm not
familiar with 40 Gb/s routers), many if not most of the current gear
handles IPv6 routing lookups in hardware, just like IPv4 (and MPLS).

For example, the mid-range platform that we use in our backbone
forwards 30 Mpps per forwarding engine, whether based on IPv4
addresses, IPv6 addresses, or MPLS labels.  30 Mpps at 1500-byte
packets corresponds to 360 Gb/s.  So, no sweat.

Routing table lookups(*) are what's most relevant here, because the other
work in forwarding is identical between IPv4 and IPv6.  Again, many
platforms are able to do line-rate forwarding between 10 Gb/s ports.
-- 
Simon, AS559.
(*) ACLs (access control lists) are also important, but again, newer
hardware can do fairly complex IPv6 ACLs at line rate.


from the academic side of the house

2007-04-24 Thread bmanning


For the first set of IPv6 records, a team from the University of Tokyo, WIDE
Project, NTT Communications, JGN2, SURFnet, CANARIE, Pacific Northwest
Gigapop and other institutions collaborated to create a network path over
30,000 kilometers in distance, crossing 6 international networks - over 3/4
the circumference of the Earth. In doing so, the team successfully
transferred data in the single and multi-stream categories at a rate of 7.67
Gbps which is equal to 230,100 terabit-meters per second (Tb-m/s).  This
record setting attempt leveraged standard TCP to achieve the new mark.

The next day, the team used a modified version of TCP to achieve an even
greater record. Using the same 30,000 km path, the network was able to
achieve a throughput of 9.08 Gbps which is equal to 272,400 Tb-m/s for both
the IPv6 multi and single stream categories. In doing so, the team surpassed
the current IPv4 records, proving that IPv6 networks are able to provide the
same, if not better, performance as IPv4.

--bill


Re: from the academic side of the house

2007-04-24 Thread Adrian Chadd

On Tue, Apr 24, 2007, [EMAIL PROTECTED] wrote:

 The next day, the team used a modified version of TCP to achieve an even
 greater record. Using the same 30,000 km path, the network was able to
 achieve a throughput of 9.08 Gbps which is equal to 272,400 Tb-m/s for both
 the IPv6 multi and single stream categories. In doing so, the team surpassed
 the current IPv4 records, proving that IPv6 networks are able to provide the
 same, if not better, performance as IPv4.

As one of the poor bastards still involved in rolling out VoIP over satellite
delivered IP at the moment, I can safely say I'm (currently) happy noone's 
trying
to push H.323 over IPv6 over these small-sized satellite links. Lord knows
we have enough trouble getting concurrent calls through 20 + 20 + 
byte overheads when the voice payload's -20- bytes.

(That said, I'd be so much happer if the current trend 'ere wasn't to -avoid-
delivering serial ports for the satellite service so we can run VoFR or PPP
w/header compression - instead being presented IP connectivity only at
either end, but you can't have everything..)



Adrian



Re: from the academic side of the house

2007-04-24 Thread Leigh Porter


Adrian Chadd wrote:

On Tue, Apr 24, 2007, [EMAIL PROTECTED] wrote:

  

The next day, the team used a modified version of TCP to achieve an even
greater record. Using the same 30,000 km path, the network was able to
achieve a throughput of 9.08 Gbps which is equal to 272,400 Tb-m/s for both
the IPv6 multi and single stream categories. In doing so, the team surpassed
the current IPv4 records, proving that IPv6 networks are able to provide the
same, if not better, performance as IPv4.



As one of the poor bastards still involved in rolling out VoIP over satellite
delivered IP at the moment, I can safely say I'm (currently) happy noone's 
trying
to push H.323 over IPv6 over these small-sized satellite links. Lord knows
we have enough trouble getting concurrent calls through 20 + 20 + 
byte overheads when the voice payload's -20- bytes.


(That said, I'd be so much happer if the current trend 'ere wasn't to -avoid-
delivering serial ports for the satellite service so we can run VoFR or PPP
w/header compression - instead being presented IP connectivity only at
either end, but you can't have everything..)



Adrian
  

Does anybody have any working v6 header suppression/compression working yet?

When I was doing VoIP over VSAT people kept trying to give me modems 
with Ethernet on them, not good for doing any header compression.


--
Leigh


Re: from the academic side of the house

2007-04-24 Thread Jim Shankland

[EMAIL PROTECTED] writes:

 The next day, the team used a modified version of TCP to achieve an
 even greater record. Using the same 30,000 km path, the network was
 able to achieve a throughput of 9.08 Gbps which is equal to 272,400
 Tb-m/s for both the IPv6 multi and single stream categories. In doing
 so, the team surpassed the current IPv4 records, proving that IPv6
 networks are able to provide the same, if not better, performance as
 IPv4.

Good job.  Two questions, though:

(1) Do the throughput figures count only the data payload (i.e.,
anything above the TCP layer), or all the bits from the protocol
stack?  If the latter, it seems a little unreasonable to credit
IPv6 with its own extra overhead -- though I'll concede that with
jumbo datagrams, that's not all that much.

(2) Getting this kind of throughput seems to depend on a fast
physical layer, plus some link-layer help (jumbo packets), plus
careful TCP tuning to deal with the large bandwidth-delay product.
The IP layer sits between the second and third of those three items.
Is there something about IPv6 vs. IPv4 that specifically improves
perfomance on this kind of test?  If so, what is it?

Jim Shankland


Re: from the academic side of the house

2007-04-24 Thread Leigh Porter


Jim Shankland wrote:

[EMAIL PROTECTED] writes:

  

The next day, the team used a modified version of TCP to achieve an
even greater record. Using the same 30,000 km path, the network was
able to achieve a throughput of 9.08 Gbps which is equal to 272,400
Tb-m/s for both the IPv6 multi and single stream categories. In doing
so, the team surpassed the current IPv4 records, proving that IPv6
networks are able to provide the same, if not better, performance as
IPv4.



Good job.  Two questions, though:

(1) Do the throughput figures count only the data payload (i.e.,
anything above the TCP layer), or all the bits from the protocol
stack?  If the latter, it seems a little unreasonable to credit
IPv6 with its own extra overhead -- though I'll concede that with
jumbo datagrams, that's not all that much.

(2) Getting this kind of throughput seems to depend on a fast
physical layer, plus some link-layer help (jumbo packets), plus
careful TCP tuning to deal with the large bandwidth-delay product.
The IP layer sits between the second and third of those three items.
Is there something about IPv6 vs. IPv4 that specifically improves
perfomance on this kind of test?  If so, what is it?

Jim Shankland
  


Also, it's a modified TCP not just tuned. I wonder how modified it is? 
Will it talk to an un-modified TCP stack (whatever that really is) ?


--
Leigh Porter