> $ easy_install hachoir-core > Searching for hachoir-core > Reading http://pypi.python.org/simple/hachoir-core/ > Reading http://hachoir.org/wiki/hachoir-core <- this page doesn't > exists anymore that's an old home url > > page, you're blocked for a while !! > > If we keep this behavior, the client-side should be more smart.
I disagree. It's the package maintainer's task to make sure the published URLs actually work. The maintainer failing to do so, I think users should be more smart and stop using an unmaintained package. Failing to do so, they should specify an explicit version. Failing to do so, they deserve waiting for the timeout. > We are adding timeout handling in Distribute, and we will probably add > a special option so it doesn't follow > external links if some distributions were found at PyPI. > > But we should find a way to remove dead links from PyPI imho. There is: ask the maintainer of the package to fix the page. > Maybe by providing a proxy for all links ? So PyPI can fallback to an > empty page if the link is dead ? I really fail to see why this is a problem. Regards, Martin _______________________________________________ Catalog-SIG mailing list [email protected] http://mail.python.org/mailman/listinfo/catalog-sig
