linklint will do exactly what you want. it's an apt-get for debian, i dont know what the status is for other distros.
regards, brett On Monday 22 November 2004 20:46, Ken Foskey wrote: > I am currently revising http://udk.openoffice.org website to fix the > wording spelling and formatting of the website. I am tripping over a > lot of dead documents in the website so I want to start with a map and > publish a list of all those documents not in the map for confirmation. > So what do I need: > > Is there a way of identifying all pages that are not linked to? > > My thought was to start with a website map that shows what links to what > and then graphviz this to show things that are orphans or branches that > are orphans. > > Any suggestions on tools that extract links from websites, website is > checked out on my hard disk so file based is fine. > > -- > Ken Foskey > OpenOffice.org -- SLUG - Sydney Linux User's Group Mailing List - http://slug.org.au/ Subscription info and FAQs: http://slug.org.au/faq/mailinglists.html