I really want a Linux-capable solution (sorry -- get cygwin if you're stuck
with windows - or better - run Virtualbox and a Linux image - free free free
;-)

I'm thinking I can get what I want by using a script to:

- do a wget on the z/VM PDF files page
- Parse between the <h4></h4>
- Parse between first <table></table> after <h4>'s
- Parse the <tr> statements to yank out the pdf URL and the human readable
title
- Do wget with -N option for the pdf files (maybe pass the script the
group(s) you want and the dir to store it in?)
- Create a  vmpdf.html file that looks like the z/VM PDF page - but contains
local links to the pdf files

I'm about 100 times better at rexx than at bash - so maybe I'll attempt this
as a rexx script and leave it to someone else to 'bash it'.   Of course,
it's all dependent on the basic layout of the PDF web page staying
consistent - so maybe an exercise in futility.. won't be the first time.

Scott

On Mon, Jul 13, 2009 at 9:35 AM, Rob van der Heij <rvdh...@gmail.com> wrote:

> On Mon, Jul 13, 2009 at 4:21 PM, Scott Rohling<scott.rohl...@gmail.com>
> wrote:
>
> > p.s.  Hmmm..  I bet I can use wget with the right incantation and get the
> > whole website to my laptop along with PDFs..   but not sure it handles
> > checking for changes?
>
> The first speed-up is to block all the surfing statistics that IBM has
> added to the various pages. I measured it to be more than 50% of the
> time it takes to load the page (whether I sit in Europe or in the US,
> no excuses there).  I modified the HTML page that lists all the books
> so that it points to the local copies of the PDFs for the ones that I
> have here handy. In fact, I even have the copy of the monitor record
> layout local on my thinkpad, but that may not be something everyone
> else needs that often ;-)
>
> And I'll do it again with next release (if any).
>
> Rob
>

Reply via email to