Erik Hetzner wrote:

Accept-Encoding is a little strange. It is used for gzip or deflate
compression, largely. I cannot imagine needing a link to a version
that is gzipped.

It is also hard to imagine why a link would want to specify the
charset to be used, possibly overriding a client’s preference. If my
browser says it can only supports UTF-8 or latin-1, it is probably
telling the truth.
Perhaps when the client/user-agent is not actually a "web browser" that is simply going to display the document to the user, but is some kind of other software. Imagine perhaps archiving software that, by policy, only will take UTF-8 encoded documents, and you need to supply a URL which is guaranteed to deliver such a thing.

Sure, the hypothetical archiving software could/should(?) just send an actual HTTP header to make sure it gets a UTF-8 charset document. But maybe sometimes it makes sense to provide an identifier that actually identifies/points to the UTF-8 charset version -- and that in the actual in-practice real world is more guaranteed to return that UTF-8 charset version from an HTTP request, without relying on content negotation which is often mis-implemented. We could probably come up with a similar reasonable-if-edge-case for encoding.

So I'm not thinking so much of "over-riding" the conneg -- I'm thinking of your initial useful framework, one URI identifies a more abstract 'document', the other identifies a specific representation. And sometimes it's probably useful to identify a specific representation in a specific charset, or, more of a stretch, encoding. No?

I notice you didn't mention 'language', I assume we agree that one is even less of a stretch, and has more clear use cases for including in a URL, like content-type.

Jonathan

Reply via email to