On Wed, Sep 26, 2012 at 4:36 PM, Shawn H Corey <shawnhco...@gmail.com> wrote: > On Wed, 26 Sep 2012 15:43:26 -0400 (EDT) > "Robert P. J. Day" <rpj...@crashcourse.ca> wrote: > >> for anyone who knows perl, i'm sure this will be trivial given the >> right perl module. if anyone can help me out, i would be grateful to >> the extent of a beer or three, or something along those lines. > > Right of the bat, I think you will need WWW::Mechanize to scrap the > website, and PDF::API2 to create the PDF. > > I don't think your project is going to be as simple as you do.
I'ld cheat: #!/bin/bash page=0 while read URL do page=$((page+1)) if [ "$page" ne "SKIP" ]; then html2ps -o - $URL | ps2pdf - $(printf page-%03d.pdf $page) fi done (warning, typed into an email client, not a shell script, so gotcha's not avoided) -- Aidan Van Dyk Create like a god, ai...@highrise.ca command like a king, http://www.highrise.ca/ work like a slave. _______________________________________________ Linux mailing list Linux@lists.oclug.on.ca http://oclug.on.ca/mailman/listinfo/linux