I really want a Linux-capable solution (sorry -- get cygwin if you're stuck with windows - or better - run Virtualbox and a Linux image - free free free ;-)
I'm thinking I can get what I want by using a script to: - do a wget on the z/VM PDF files page - Parse between the <h4></h4> - Parse between first <table></table> after <h4>'s - Parse the <tr> statements to yank out the pdf URL and the human readable title - Do wget with -N option for the pdf files (maybe pass the script the group(s) you want and the dir to store it in?) - Create a vmpdf.html file that looks like the z/VM PDF page - but contains local links to the pdf files I'm about 100 times better at rexx than at bash - so maybe I'll attempt this as a rexx script and leave it to someone else to 'bash it'. Of course, it's all dependent on the basic layout of the PDF web page staying consistent - so maybe an exercise in futility.. won't be the first time. Scott On Mon, Jul 13, 2009 at 9:35 AM, Rob van der Heij <rvdh...@gmail.com> wrote: > On Mon, Jul 13, 2009 at 4:21 PM, Scott Rohling<scott.rohl...@gmail.com> > wrote: > > > p.s. Hmmm.. I bet I can use wget with the right incantation and get the > > whole website to my laptop along with PDFs.. but not sure it handles > > checking for changes? > > The first speed-up is to block all the surfing statistics that IBM has > added to the various pages. I measured it to be more than 50% of the > time it takes to load the page (whether I sit in Europe or in the US, > no excuses there). I modified the HTML page that lists all the books > so that it points to the local copies of the PDFs for the ones that I > have here handy. In fact, I even have the copy of the monitor record > layout local on my thinkpad, but that may not be something everyone > else needs that often ;-) > > And I'll do it again with next release (if any). > > Rob >