On Fri, 2013-03-15 at 09:11 -0400, Dale H. Cook wrote:
> At 09:44 PM 3/14/2013, tamouse mailing lists wrote:
>
> >If you are delivering files to a (human) user via their browser, by whatever
> >mechanism, that means someone can write a script to scrape them.
>
> That script, however, would have
At 09:44 PM 3/14/2013, tamouse mailing lists wrote:
>If you are delivering files to a (human) user via their browser, by whatever
>mechanism, that means someone can write a script to scrape them.
That script, however, would have to be running on my host system in order to
access the script whic
At 04:06 AM 3/14/2013, tamouse mailing lists wrote:
>If the files are delivered via the web, by php or some other means, even if
>located outside webroot, they'd still be scrapeable.
Bots, however, being "mechanical" (i.e., hard wired or programmed) behave in
different ways than humans, and that
On Mar 13, 2013 7:06 PM, "David Robley" wrote:
>
> "Dale H. Cook" wrote:
>
> > At 05:04 PM 3/13/2013, Dan McCullough wrote
> > :
> >>Web bots can ignore the robots.txt file, most scrapers would.
> >
> > and at 05:06 PM 3/13/2013, Marc Guay wrote:
> >
> >>These don't sound like robots that would re
"Dale H. Cook" wrote:
> At 05:04 PM 3/13/2013, Dan McCullough wrote
> :
>>Web bots can ignore the robots.txt file, most scrapers would.
>
> and at 05:06 PM 3/13/2013, Marc Guay wrote:
>
>>These don't sound like robots that would respect a txt file to me.
>
> Dan and Marc are correct. Although I
At 05:04 PM 3/13/2013, Dan McCullough wrote
:
>Web bots can ignore the robots.txt file, most scrapers would.
and at 05:06 PM 3/13/2013, Marc Guay wrote:
>These don't sound like robots that would respect a txt file to me.
Dan and Marc are correct. Although I used the terms "spiders" and "pirates"
At 04:58 PM 3/13/2013, Jen Rasmussen wrote:
>Have you tried keeping all of your documents in one directory and blocking
>that directory via a robots.txt file?
A spider used by a pirate site does not have to honor robots.txt, just as a
non-Adobe PDF utility does not have to honor security setting
> Have you tried keeping all of your documents in one directory and blocking
> that directory via a robots.txt file?
These don't sound like robots that would respect a txt file to me.
--
PHP General Mailing List (http://www.php.net/)
To unsubscribe, visit: http://www.php.net/unsub.php
Web bots can ignore the robots.txt file, most scrapers would.
On Mar 13, 2013 4:59 PM, "Jen Rasmussen" wrote:
> -Original Message-
> From: Dale H. Cook [mailto:radiot...@plymouthcolony.net]
> Sent: Wednesday, March 13, 2013 3:38 PM
> To: php-general@lists.php.net
> Subject: [PHP] Accessin
-Original Message-
From: Dale H. Cook [mailto:radiot...@plymouthcolony.net]
Sent: Wednesday, March 13, 2013 3:38 PM
To: php-general@lists.php.net
Subject: [PHP] Accessing Files Outside the Web Root
Let me preface my question by noting that I am virtually a PHP novice.
Although I am a long
10 matches
Mail list logo