RE: Extracting Images Question

2004-11-08 Thread Thomas, Mark - BLS CTR
> I wish it was that easy, I could have a script ftp and grab 
> them. Our company is very funny and does not work well 
> between the different depts, so getting ftp access to a 
> directory on another computer on the other side of the world 
> won't happen. How hard is it to do what I stated in my original post?

It's easy with WWW::Mechanize, as stated in an earlier response.

-- 
Mark Thomas 
Internet Systems Architect
___
BAE SYSTEMS Information Technology 
2525 Network Place
Herndon, VA  20171  USA 

___
Perl-Win32-Users mailing list
[EMAIL PROTECTED]
To unsubscribe: http://listserv.ActiveState.com/mailman/mysubs


Re: Extracting Images Question

2004-11-06 Thread $Bill Luebkert
Allan wrote:

> I wish it was that easy, I could have a script ftp and grab them. Our
> company is very funny and does not work well between the different depts, so
> getting ftp access to a directory on another computer on the other side of
> the world won't happen. How hard is it to do what I stated in my original
> post?

Not too difficult.  It might be easier if you had directory listing
access to the directory the files are in.  In Apache, that would be
something like:


Options Indexes
IndexOptions FancyIndexing
Order Deny,Allow
Allow from All


This could also be password controlled via .htpasswd file.

AuthUserFile /path/to/pics/.htpasswd
AuthGroupFile /dev/null
AuthName "Protected Files"
AuthType Basic

And then 'Require valid-user' could be added to the directory indexing.

Without indexing, you would have to parse the page(s) for links to pics
and process them for new ? files or whatever criteria you are using to
detemine whether to D/L.

FTP access to a directory that contains only files that you could
D/L via HTTP anyway shouldn't be any sort of problem security wise.

-- 
  ,-/-  __  _  _ $Bill LuebkertMailto:[EMAIL PROTECTED]
 (_/   /  )// //   DBE CollectiblesMailto:[EMAIL PROTECTED]
  / ) /--<  o // //  Castle of Medieval Myth & Magic http://www.todbe.com/
-/-' /___/_<_http://dbecoll.tripod.com/ (My Perl/Lakers stuff)
___
Perl-Win32-Users mailing list
[EMAIL PROTECTED]
To unsubscribe: http://listserv.ActiveState.com/mailman/mysubs


RE: Extracting Images Question

2004-11-06 Thread Allan
I wish it was that easy, I could have a script ftp and grab them. Our
company is very funny and does not work well between the different depts, so
getting ftp access to a directory on another computer on the other side of
the world won't happen. How hard is it to do what I stated in my original
post?

Thank you
Allan

-Original Message-
From: $Bill Luebkert [mailto:[EMAIL PROTECTED]
Sent: Saturday, November 06, 2004 1:48 AM
To: [EMAIL PROTECTED]
Cc: [EMAIL PROTECTED]
Subject: Re: Extracting Images Question


[EMAIL PROTECTED] wrote:

> I know there has to be an easy Perl way to do this. Our HR
> dept. scans in thousands of documents and has them up on a
> intranet site with thumbnail views. Then another department
> clicks on the thumbnail to get the large version of that
> doc to save to their desktop. How can I use Perl to enter
> in the site name http://myintranet.com and say this hour
> there are 40 new scan images up there it will parse through
> that page and follow the thumbnail nail paths and download
> all the actual size images.
>
> I can use LWP but this seems a bit daunting. Then I would
> need to pattern match and so on.
>
> Any ideas?

If you have FTP access, you could handle it a lot more easily
(by checking file times etc.) and then just download files
that are of different dates or possibly sizes than previous
versions of files.

You would need to be more exact in how the files differ and
are added to be more specific in a solution.  Can you keep
a local mirror ?  That would make it even easier since you
can just compare what you have locally to what's new on the
site.

--
  ,-/-  __  _  _ $Bill LuebkertMailto:[EMAIL PROTECTED]
 (_/   /  )// //   DBE CollectiblesMailto:[EMAIL PROTECTED]
  / ) /--<  o // //  Castle of Medieval Myth & Magic
http://www.todbe.com/
-/-' /___/_<_http://dbecoll.tripod.com/ (My Perl/Lakers stuff)

___
Perl-Win32-Users mailing list
[EMAIL PROTECTED]
To unsubscribe: http://listserv.ActiveState.com/mailman/mysubs


Re: Extracting Images Question

2004-11-05 Thread $Bill Luebkert
[EMAIL PROTECTED] wrote:

> I know there has to be an easy Perl way to do this. Our HR
> dept. scans in thousands of documents and has them up on a
> intranet site with thumbnail views. Then another department
> clicks on the thumbnail to get the large version of that
> doc to save to their desktop. How can I use Perl to enter
> in the site name http://myintranet.com and say this hour
> there are 40 new scan images up there it will parse through
> that page and follow the thumbnail nail paths and download
> all the actual size images.
> 
> I can use LWP but this seems a bit daunting. Then I would
> need to pattern match and so on.
> 
> Any ideas?

If you have FTP access, you could handle it a lot more easily
(by checking file times etc.) and then just download files
that are of different dates or possibly sizes than previous
versions of files.

You would need to be more exact in how the files differ and
are added to be more specific in a solution.  Can you keep
a local mirror ?  That would make it even easier since you
can just compare what you have locally to what's new on the
site.

-- 
  ,-/-  __  _  _ $Bill LuebkertMailto:[EMAIL PROTECTED]
 (_/   /  )// //   DBE CollectiblesMailto:[EMAIL PROTECTED]
  / ) /--<  o // //  Castle of Medieval Myth & Magic http://www.todbe.com/
-/-' /___/_<_http://dbecoll.tripod.com/ (My Perl/Lakers stuff)
___
Perl-Win32-Users mailing list
[EMAIL PROTECTED]
To unsubscribe: http://listserv.ActiveState.com/mailman/mysubs


Re: Extracting Images Question

2004-11-05 Thread Mike G.
check out saime or mechanize.  Both have the ability to follow links in a page and 
then return to the original window.  How well they work depends upon how the web page 
is constructed.

Mike

On Fri, Nov 05, 2004 at 10:36:27PM -0500, [EMAIL PROTECTED] wrote:
> I know there has to be an easy Perl way to do this. Our HR
> dept. scans in thousands of documents and has them up on a
> intranet site with thumbnail views. Then another department
> clicks on the thumbnail to get the large version of that
> doc to save to their desktop. How can I use Perl to enter
> in the site name http://myintranet.com and say this hour
> there are 40 new scan images up there it will parse through
> that page and follow the thumbnail nail paths and download
> all the actual size images.
> 
> I can use LWP but this seems a bit daunting. Then I would
> need to pattern match and so on.
> 
> Any ideas?
> 
> Thank you
> Allan
> ___
> Perl-Win32-Users mailing list
> [EMAIL PROTECTED]
> To unsubscribe: http://listserv.ActiveState.com/mailman/mysubs
___
Perl-Win32-Users mailing list
[EMAIL PROTECTED]
To unsubscribe: http://listserv.ActiveState.com/mailman/mysubs


Extracting Images Question

2004-11-05 Thread barons
I know there has to be an easy Perl way to do this. Our HR
dept. scans in thousands of documents and has them up on a
intranet site with thumbnail views. Then another department
clicks on the thumbnail to get the large version of that
doc to save to their desktop. How can I use Perl to enter
in the site name http://myintranet.com and say this hour
there are 40 new scan images up there it will parse through
that page and follow the thumbnail nail paths and download
all the actual size images.

I can use LWP but this seems a bit daunting. Then I would
need to pattern match and so on.

Any ideas?

Thank you
Allan
___
Perl-Win32-Users mailing list
[EMAIL PROTECTED]
To unsubscribe: http://listserv.ActiveState.com/mailman/mysubs