On Wed, 2011-01-19 at 10:58 -0500, Mike Rathburn wrote:
> You can speed up and control in more granular detail the Google stuff
> by signing up an account with Google Webmaster Tools.
> 

Also with regard to the Crawler Access portion of the Google Webmasters
tools. You are not provided with anything to speed up, much less control
in a granular way crawler access. Its a lame tool for generating
robots.txt files, IMHO. Here is an example I just generated, but via
hand would have been much faster.

User-agent: *
Allow: /

User-agent: Mediapartners-Google
Disallow: /

User-agent: Googlebot
Disallow: /

User-agent: Adsbot-Google
Disallow: /

User-agent: Googlebot-Image
Disallow: /

User-agent: Googlebot-Mobile
Disallow: /

However there is no mention if robots.txt files are parsed in order. The
first entry might negate all others, and the rest be ignored. If it was
like the a firewall, you would do allow after deny. But gain this is
straight from Googles own tool in their Webmasters tool.

Who really needs a robots.txt file generator. The one Google provides
does nothing to help out wrt to the problems on the wiki. That has to be
done that manually. What Googles tool does most anyone can do by hand
just the same in less time.


Now it was useful for the following

Line 3: Crawl-delay: 15 Rule ignored by Googlebot

-- 
William L. Thomson Jr.
Obsidian-Studios, Inc.
http://www.obsidian-studios.com


---------------------------------------------------------------------
Archive      http://marc.info/?l=jaxlug-list&r=1&w=2
RSS Feed     http://www.mail-archive.com/[email protected]/maillist.xml
Unsubscribe  [email protected]

Reply via email to