Just got a copy of Plucker last week and it's been a great help to me (I switched from AvantGo, when I find it refused to download some of the 200k-300k files I wanted to read offline).
I was just wondering if there is anyone working on putting some persistent caching into the python scripts? I find on my (windows + 56k modem) machine it can take quite a few min to update the hundred or so individual pages I end up synching with (not forgetting the amount of wasted network bandwidth). If no one else is doing it I might teach myself Python/http and add it myself. On a related note, does anyone else think a 'MAXLINKS' feature would be good? Several of the pages I am interested in have a history page which makes it easy to download all the latest updates, but obviously a whole years worth of page history is rather a lot, and so some way of grabbing only the first N links could be helpful (currently I've hacked my local scripts to stop at 15, but it's not very configurable). Davo
