I use SiteScooper as the front end to plucking all my sites.  It caches them
and can even pre-process the html files to rid them of banners, etc.

It is at www.sitescooper.org but beware - the learning curve is pretty
steep.  There are quite a few sample .site files available though.  Good
luck.

-----Original Message-----
From: Davo [mailto:[EMAIL PROTECTED]]
Sent: Monday, January 21, 2002 2:46 PM
To: Plucker Development
Subject: Hotsynch persistent caching?


Just got a copy of Plucker last week and it's been a great help to me (I
switched from AvantGo, when I find it refused to download some of the
200k-300k files I wanted to read offline).

I was just wondering if there is anyone working on putting some persistent
caching into the python scripts? I find on my (windows + 56k modem) machine
it can take quite a few min to update the hundred or so individual pages I
end up synching with (not forgetting the amount of wasted network
bandwidth). If no one else is doing it I might teach myself Python/http and
add it myself.

On a related note, does anyone else think a 'MAXLINKS' feature would be
good? Several of the pages I am interested in have a history page which
makes it easy to download all the latest updates, but obviously a whole
years worth of page history is rather a lot, and so some way of grabbing
only the first N links could be helpful (currently I've hacked my local
scripts to stop at 15, but it's not very configurable).

Davo



Reply via email to