Considering that one of your core use cases for this is security-
related, I'm surprised that you're effectively arguing that HTTP and
HTTPS URLs with the same authority be collapsed into one name space.
Many standards and common practices currently sandbox policy and
metadata to a single URL scheme + authority by default, including
robots.txt, p3p.xml, cookie scoping, automated redirection processing
in HTTP, cache invalidation, OPTIONS metadata, cross-site scripting
and I'm sure quite a few more. This is the de facto standard for what
a "Web site" is, and while there are many other colloquial meanings of
that phrase, this is current technical practice.
Trying to establish a standard for site-wide metadata that doesn't
follow this practice is IMO doomed to sow yet more confusion about an
already muddled area, and potentially open up security as well as
usability and technical problems.
That said, there's nothing to stop a particular application -- e.g.,
OpenID -- saying that for a particular purpose, site-meta should be
checked on a HTTP URL even though the URL presented is mailto: (for
example), or even that www.example.com should be tried if example.com
isn't available (although I still don't think it's necessary).
What I'm not willing to do is enshrine these things in standards that
are supposed to help extend the Web architecture, not dilute it. The
fact that a few $2 Web hosts don't provide adequate control to their
customers in 2008 should not affect something so fundamental as the
definition of what a Web site is for the next 30 years (if this
succeeds, of course).
Cheers,
On 03/12/2008, at 6:32 PM, Eran Hammer-Lahav wrote:
On 02/12/2008, at 4:24 PM, Mark Nottingham wrote:
/site-meta on http://foobar.com/ doesn't (and can't, on its own) make
any authoritative assertions about mailto:[EMAIL PROTECTED]; even
though
the authority is the same, the URI scheme is different.
I know this particular issue is an important one to the OpenID folks,
but there needs to be a very careful and broad discussion of allowing
policy and metadata from HTTP to be considered *automatically*
authoritative for other protocols.
I do not considered /site-meta to be about HTTP resources. It is
metadata about the domain authority and uses HTTP as the protocol to
deliver that document. It can equally link to HTTP URIs as to other
URIs (i.e. point to its robots.txt available at an ftp:// URI). I
think it is safe to assume that whoever controls the domain controls
any URI scheme within that domain. Companies can split control
between departments but you go high enough there is one entity which
owns everything under that authority.
HTTP clearly allows: 'GET mailto:[EMAIL PROTECTED]', but what is
actually served is up to the server. In theory, that can serve a 303
with Link header to the XRD describing the identifier. The problem,
of course, is that most web servers will fail on such request, or at
least most platforms will not allow the developer easy access to
control the response to such requests. But the point is, the HTTP
protocol is nowhere restricted to provide information about HTTP
URIs alone. The fact that user-agents will use HTTP when the URI
scheme is HTTP and use FTP when the URI scheme is FTP is more of a
practical convention than a strict requirement.
The issue of what constitute authoritative metadata with regard to
the domain authority is not something we can resolve beyond the
reasonable expectation that the entity that control the domain has
sufficient authority. Can the profile.yahoo.com admin be considered
the authority for my profile page? In the context of discovery, I
believe the answer is yes. Philosophically, I can argue that only
the profile owner has the authority to control that page, but such
control in today's infrastructure, is eventually enforced by the
domain admin anyway.
EHL
--
Mark Nottingham http://www.mnot.net/