RE: Extracting Images Question
> I wish it was that easy, I could have a script ftp and grab > them. Our company is very funny and does not work well > between the different depts, so getting ftp access to a > directory on another computer on the other side of the world > won't happen. How hard is it to do what I stated in my original post? It's easy with WWW::Mechanize, as stated in an earlier response. -- Mark Thomas Internet Systems Architect ___ BAE SYSTEMS Information Technology 2525 Network Place Herndon, VA 20171 USA ___ Perl-Win32-Users mailing list [EMAIL PROTECTED] To unsubscribe: http://listserv.ActiveState.com/mailman/mysubs
Re: Extracting Images Question
Allan wrote: > I wish it was that easy, I could have a script ftp and grab them. Our > company is very funny and does not work well between the different depts, so > getting ftp access to a directory on another computer on the other side of > the world won't happen. How hard is it to do what I stated in my original > post? Not too difficult. It might be easier if you had directory listing access to the directory the files are in. In Apache, that would be something like: Options Indexes IndexOptions FancyIndexing Order Deny,Allow Allow from All This could also be password controlled via .htpasswd file. AuthUserFile /path/to/pics/.htpasswd AuthGroupFile /dev/null AuthName "Protected Files" AuthType Basic And then 'Require valid-user' could be added to the directory indexing. Without indexing, you would have to parse the page(s) for links to pics and process them for new ? files or whatever criteria you are using to detemine whether to D/L. FTP access to a directory that contains only files that you could D/L via HTTP anyway shouldn't be any sort of problem security wise. -- ,-/- __ _ _ $Bill LuebkertMailto:[EMAIL PROTECTED] (_/ / )// // DBE CollectiblesMailto:[EMAIL PROTECTED] / ) /--< o // // Castle of Medieval Myth & Magic http://www.todbe.com/ -/-' /___/_<_http://dbecoll.tripod.com/ (My Perl/Lakers stuff) ___ Perl-Win32-Users mailing list [EMAIL PROTECTED] To unsubscribe: http://listserv.ActiveState.com/mailman/mysubs
RE: Extracting Images Question
I wish it was that easy, I could have a script ftp and grab them. Our company is very funny and does not work well between the different depts, so getting ftp access to a directory on another computer on the other side of the world won't happen. How hard is it to do what I stated in my original post? Thank you Allan -Original Message- From: $Bill Luebkert [mailto:[EMAIL PROTECTED] Sent: Saturday, November 06, 2004 1:48 AM To: [EMAIL PROTECTED] Cc: [EMAIL PROTECTED] Subject: Re: Extracting Images Question [EMAIL PROTECTED] wrote: > I know there has to be an easy Perl way to do this. Our HR > dept. scans in thousands of documents and has them up on a > intranet site with thumbnail views. Then another department > clicks on the thumbnail to get the large version of that > doc to save to their desktop. How can I use Perl to enter > in the site name http://myintranet.com and say this hour > there are 40 new scan images up there it will parse through > that page and follow the thumbnail nail paths and download > all the actual size images. > > I can use LWP but this seems a bit daunting. Then I would > need to pattern match and so on. > > Any ideas? If you have FTP access, you could handle it a lot more easily (by checking file times etc.) and then just download files that are of different dates or possibly sizes than previous versions of files. You would need to be more exact in how the files differ and are added to be more specific in a solution. Can you keep a local mirror ? That would make it even easier since you can just compare what you have locally to what's new on the site. -- ,-/- __ _ _ $Bill LuebkertMailto:[EMAIL PROTECTED] (_/ / )// // DBE CollectiblesMailto:[EMAIL PROTECTED] / ) /--< o // // Castle of Medieval Myth & Magic http://www.todbe.com/ -/-' /___/_<_http://dbecoll.tripod.com/ (My Perl/Lakers stuff) ___ Perl-Win32-Users mailing list [EMAIL PROTECTED] To unsubscribe: http://listserv.ActiveState.com/mailman/mysubs
Re: Extracting Images Question
[EMAIL PROTECTED] wrote: > I know there has to be an easy Perl way to do this. Our HR > dept. scans in thousands of documents and has them up on a > intranet site with thumbnail views. Then another department > clicks on the thumbnail to get the large version of that > doc to save to their desktop. How can I use Perl to enter > in the site name http://myintranet.com and say this hour > there are 40 new scan images up there it will parse through > that page and follow the thumbnail nail paths and download > all the actual size images. > > I can use LWP but this seems a bit daunting. Then I would > need to pattern match and so on. > > Any ideas? If you have FTP access, you could handle it a lot more easily (by checking file times etc.) and then just download files that are of different dates or possibly sizes than previous versions of files. You would need to be more exact in how the files differ and are added to be more specific in a solution. Can you keep a local mirror ? That would make it even easier since you can just compare what you have locally to what's new on the site. -- ,-/- __ _ _ $Bill LuebkertMailto:[EMAIL PROTECTED] (_/ / )// // DBE CollectiblesMailto:[EMAIL PROTECTED] / ) /--< o // // Castle of Medieval Myth & Magic http://www.todbe.com/ -/-' /___/_<_http://dbecoll.tripod.com/ (My Perl/Lakers stuff) ___ Perl-Win32-Users mailing list [EMAIL PROTECTED] To unsubscribe: http://listserv.ActiveState.com/mailman/mysubs
Re: Extracting Images Question
check out saime or mechanize. Both have the ability to follow links in a page and then return to the original window. How well they work depends upon how the web page is constructed. Mike On Fri, Nov 05, 2004 at 10:36:27PM -0500, [EMAIL PROTECTED] wrote: > I know there has to be an easy Perl way to do this. Our HR > dept. scans in thousands of documents and has them up on a > intranet site with thumbnail views. Then another department > clicks on the thumbnail to get the large version of that > doc to save to their desktop. How can I use Perl to enter > in the site name http://myintranet.com and say this hour > there are 40 new scan images up there it will parse through > that page and follow the thumbnail nail paths and download > all the actual size images. > > I can use LWP but this seems a bit daunting. Then I would > need to pattern match and so on. > > Any ideas? > > Thank you > Allan > ___ > Perl-Win32-Users mailing list > [EMAIL PROTECTED] > To unsubscribe: http://listserv.ActiveState.com/mailman/mysubs ___ Perl-Win32-Users mailing list [EMAIL PROTECTED] To unsubscribe: http://listserv.ActiveState.com/mailman/mysubs
Extracting Images Question
I know there has to be an easy Perl way to do this. Our HR dept. scans in thousands of documents and has them up on a intranet site with thumbnail views. Then another department clicks on the thumbnail to get the large version of that doc to save to their desktop. How can I use Perl to enter in the site name http://myintranet.com and say this hour there are 40 new scan images up there it will parse through that page and follow the thumbnail nail paths and download all the actual size images. I can use LWP but this seems a bit daunting. Then I would need to pattern match and so on. Any ideas? Thank you Allan ___ Perl-Win32-Users mailing list [EMAIL PROTECTED] To unsubscribe: http://listserv.ActiveState.com/mailman/mysubs