On Thursday, August 25, 2005, at 03:12  PM, Walter Underwood wrote:
I would call desktop clients "clients" not "robots". The distinction is
how they add feeds to the polling list. Clients add them because of
human decisions. Robots discover them mechanically and add them.

So, clients should act like browsers, and ignore robots.txt.

How could this all be related to aggregators that accept feed URL submissions? I'd imagine the desired behavior is the same as for crawlers--should they check for robots.txt at the root of any domain where a feed is submitted? How about cases where the feed is hosted on a site other than the website that it's tied to (for example, a service like FeedBurner) so some other site's robot.txt controls access to the feed (...or at least tries to)?

We've already rejected the idea of trying to build DRM into feeds--is there some way to sidestep the legal complexities and problems that would arise from trying to to that and at the same time enable machine readable statements about what the publisher wants to allow others to do with the feed, and things they want to prohibit, into the feed? If we're not qualified to design an extension to do that, is there someone else who is qualified, and who cares enough to do it?

Reply via email to