I think you're misunderstanding what was supposed to happen.

The robots.txt file is only followed for links that wget is
automatically following. This means (a) wget has to be in
recursive-descent mode (-r or -m), and (b) it only applies to links that
weren't explicitly requested by the user. In other words, it applies
only to links that wget is actually robotting on.

Hope that helps.

-mjc

On 03/16/2012 01:04 PM, phil curb wrote:
> I just tried creating a web server locally.
> |I tried creating a web server locally  putting robots.txt in there and using 
> wget  and it didn't work
> 
> 
> 
> http://pastebin.com/raw.php?i=kt1mV2af 
> 
> 
> C:\r>wget 127.0.0.1:56
> ....
> 2012-03-16 19:45:32 (20.0 KB/s) - `index.html' saved [3/3] C:\r>wget 
> 127.0.0.1:56/robots.txt
> ....
> 2012-03-16 19:45:43 (175 KB/s) - `robots.txt' saved [26/26] C:\r>type 
> robots.txt
> User-agent: *
> Disallow: /
> C:\r>


Reply via email to