I use SiteScooper as the front end to plucking all my sites. It caches them and can even pre-process the html files to rid them of banners, etc.
It is at www.sitescooper.org but beware - the learning curve is pretty steep. There are quite a few sample .site files available though. Good luck. -----Original Message----- From: Davo [mailto:[EMAIL PROTECTED]] Sent: Monday, January 21, 2002 2:46 PM To: Plucker Development Subject: Hotsynch persistent caching? Just got a copy of Plucker last week and it's been a great help to me (I switched from AvantGo, when I find it refused to download some of the 200k-300k files I wanted to read offline). I was just wondering if there is anyone working on putting some persistent caching into the python scripts? I find on my (windows + 56k modem) machine it can take quite a few min to update the hundred or so individual pages I end up synching with (not forgetting the amount of wasted network bandwidth). If no one else is doing it I might teach myself Python/http and add it myself. On a related note, does anyone else think a 'MAXLINKS' feature would be good? Several of the pages I am interested in have a history page which makes it easy to download all the latest updates, but obviously a whole years worth of page history is rather a lot, and so some way of grabbing only the first N links could be helpful (currently I've hacked my local scripts to stop at 15, but it's not very configurable). Davo
