More below.

On 12/05/2017, 16:27, Michael Welzl wrote:
On May 12, 2017, at 3:24 PM, Gorry Fairhurst<go...@erg.abdn.ac.uk>  wrote:

See below.

On 12/05/2017, 13:31, Michael Welzl wrote:
Hi,

Thanks a lot for all your comments (plus the nits we authors of the other 
-usage draft received offline).

I’ll try to address them all - but there are a two technical questions in this 
email that made me stop, so I’ll cut all the editorial stuff away and discuss 
them here - in line below:


- Why do this??? - Isn't it better to set flow labels per interface or for the 
whole stack, how can any specific transport or application pick unique labels?
TEXT:
   o  Specify IPv6 flow label field
      Protocols: SCTP
(i.e., Is this automatable by the host and a host wide
configuration?)
Somehow the question seems irrelevant in the context of this draft, which is a 
list of transport features of protocols. These features are defined in the RFCs 
spec’ing the protocols - for SCTP, this is defined, and that’s why it’s here.

We can discuss this for the proposed services that a system should offer, which 
we try to write up in the minset draft:
I do think that an application should be allowed to assign a label to a TAPS 
flow (as we call them), which could then map to this function. I mean, isn’t a 
flow label supposed to identify a transport flow? Then a system-wide 
configuration wouldn't seem right to me.
I think we may disagree. Flow ids identify flows to the network layer, they 
have no role at the transport layer, and need to be unique (as best they can) 
for a source address.
We disagree indeed - in particular about the “unique (as best they can)..” bit. 
Where is this written??
I'm taking the position of using this as input to an ECMP or LAG hash algorithm.

I much prefer the idea that the Flow id is generated by the IP system, by using 
a hash - possibly utilising transport data as a part of this hash, and 
including the protocol number.
RFC 6437 introduces the flow label as a replacement for the 5-tuple - “possibly 
utilising transport data as a part of this hash” seems to me to be a very weak 
requirement here!
OK, which I think is the idea in RFC6438.
Anyway classifiers in the network wouldn’t work on the flow label alone, but, 
from RFC 6437, section 2 which is called “specification”:
   "Packet classifiers can
    use the triplet of Flow Label, Source Address, and Destination
    Address fields to identify the flow to which a particular packet
    belongs."
Yes.
Then what is the flow label good for, if it’s unique per source address? It 
doesn’t add any information to this 3-tuple in this case!
Aha - I mean each "microflow" sent from a specific source address should be identified by a different and unique flow ID.
That seems to be what ECMP is expecting and I suspect ECMP is an improtant 
use-case.

The alternative (if I understand) could be: I could imagine each application 
could (in theory) be provided with an API to find out what flow-ids are 
currently being used for each interface it cares about and to then reserve one 
of the unused IDs for the specific interface(s) that it wishes to use. Then we 
need to ensure all upper layer entities coordinate on this. To me, this seems 
over-kill, and the approach taken with ephemeral port assignment is much 
simpler - the application simply doesn't get involved with choosing the number.

Now if what you are saying is that you want the App to somehow signal that it 
can use an existing flow ID that is in use, and combine data with that flow to 
get the same network treeatment, I can understand the case. However, that's not 
exactly the same thing.
I understand that it would be nice to avoid upper-layer coordination here. 
However, I see at least two use cases for the application being more in control:
1) avoiding fate sharing (encouraging ECMP), e.g. for increased resilience
Yes. Part of the idea here is that microflows (say with the same IPsec ESP) can now be separately forwarded if that is what is desired by the sending endpoint.
2) the opposite: grouping flows, to be able to apply priorities on them, using 
a mechanism such as the Congestion Manager or 
https://tools.ietf.org/html/draft-welzl-tcp-ccc
That's the converse of the IPsec ESP example above, and also ok if the endpoint wishes this.
So this is not about giving the application control of the specific flow label 
number, but allowing it to say “use the same number for these flows” or not.
That's fine with me. Providing it is *NOT* the flow-id, but an input to the function that determines the flow-id.
I think this could nicely be done by letting it number flows, and grouping them 
via equal numbering - without guaranteeing that these numbers map onto the 
exact same numbers as a flow label.
OK.

-------------------
Get Interface MTU is missing from pass 2 and 3:

ADD to pass 2:

        GET_INTERFACE_MTU.UDP:
                Pass 1 primitive: GET_INTERFACE_MTU
                Returns: Maximum datagram size (bytes)
But this doesn’t exist!
I think I don't understand your comment ... and interpretting low-numbered RFCs 
is never easy -  I'll use RFC1122 as my basis:

RFC 1122 says:
       " A host MUST implement a mechanism to allow the transport layer
         to learn MMS_S, the maximum transport-layer message size that
         may be sent for a given {source, destination, TOS} triplet..."
       " and EMTU_S must be less than or equal to the MTU of the network
         interface corresponding to the source address of the datagram."

TCP handles this for the app.
… and UDP is another such transport layer. If you try to send a message that’s 
too large, it throws an error, based on the information it gets via the 
paragraph you quote above.
But that’s UDP, not the application on top.
I don't expect each UDP app to have to start by trying to send 64,000B and reducing to see what works. Apps need to be able to find this out.

(If you happened to implement host fragementation, any size of send upto your fragmentation limit would always return true when writing).

  It’s strictly an IP function and I couldn’t find it described for UDP 
anywhere. I think we agreed on how a TAPS system should handle this, and this 
is reflected in
https://tools.ietf.org/html/draft-gjessing-taps-minset-04#section-6.4.1
… which may require a system to implement new local functionality, maybe based 
on this MTU function - but to my understanding it’s just not something that UDP 
offers.
It's something that a UDP App really needs to pay attention to as per RFC8085 - we may 
differ on whether you call that "offers" or needs to function. Either way, an 
app that plans to use any form of PMTUD needs to use this number.
I agree; and we have put related functions into the minset draft.
Yes we do agree. If you want to redefine that to bytes permitted in the UDP payload, I would also be really happy.
But here we’re describing what it is that UDP itself (not a full-fleged TAPS 
system) currently offers…

And for that, the UDP app must assume either the network-wide default, or base this on maths from the MTU info at the network-layer (section 3.2 of RFC8085). That's part of using UDP.
As put in RFC1122:
       " A host that does not implement local fragmentation MUST ensure
         that the transport layer (for TCP) or the application layer
         (for UDP) obtains MMS_S from the IP layer and does not send a
         datagram exceeding MMS_S in size.”
Okaaay, you found an instance of  “ _ or the application layer (for UDP) _ “.   
 I agree this should be included!  But, this is not the interface MTU.
Maybe it's better to read PMTU, and I'd be fine with a socket (or whatever API) retrieving that in place of the Interface-MTU).
 From RFC 1122:

***
          If no local fragmentation
          is performed, the value of MMS_S will be:

             MMS_S = EMTU_S -<IP header size>

          and EMTU_S must be less than or equal to the MTU of the network
          interface corresponding to the source address of the datagram. 
paragraph further above defines MMS_S as the maximum transport-layer m
***

So first of all, there’s the “if” - this would only be the value without local 
fragmentation. Can we assume that there won’t be local fragmentation?
We put the discussion text of host fragmentation in the "DF" discussion of our draft.

You could argue that we separate these because "DF" can be set on host fragments. (IPv6
Then, EMTU_S<= interface MTU,
OK
and MMS_S is even smaller: very reasonably, the IP header size is subtracted.
OK, I'd also be happy to say retrieves MMS_S. (I'm not sure though this in practice is implemented ?) -- but does this primitive exist in the real world? or do we need to explain both?
Cheers,
Michael

_______________________________________________
Taps mailing list
Taps@ietf.org
https://www.ietf.org/mailman/listinfo/taps

Gorry

_______________________________________________
Taps mailing list
Taps@ietf.org
https://www.ietf.org/mailman/listinfo/taps

Reply via email to