Just got a copy of Plucker last week and it's been a great help to me (I
switched from AvantGo, when I find it refused to download some of the
200k-300k files I wanted to read offline).

I was just wondering if there is anyone working on putting some persistent
caching into the python scripts? I find on my (windows + 56k modem) machine
it can take quite a few min to update the hundred or so individual pages I
end up synching with (not forgetting the amount of wasted network
bandwidth). If no one else is doing it I might teach myself Python/http and
add it myself.

On a related note, does anyone else think a 'MAXLINKS' feature would be
good? Several of the pages I am interested in have a history page which
makes it easy to download all the latest updates, but obviously a whole
years worth of page history is rather a lot, and so some way of grabbing
only the first N links could be helpful (currently I've hacked my local
scripts to stop at 15, but it's not very configurable).

Davo


Reply via email to