Bobby Impollonia wrote: > > Rereading what you posted, by "such a prominent place on Google", did > you mean specifically the ".4" and ".3" links that show up below > sqlalchemy when www.sqlalchemy.org is returned in the search results? > Those are what google calls "sitelinks". You can tell them not to use > certain pages as sitelinks via the google webmaster tools. They'll > remove them, but they don't get replaced so you will have two fewer > sitelinks then.
I did mean that, but im fine just delisting them totally. > > On Thu, May 21, 2009 at 8:03 PM, Bobby Impollonia <bob...@gmail.com> > wrote: >>> otherwise if you have any advice on how to get 0.4/0.3 >>> delisted from such a prominent place on Google, that would be >>> appreciated. >> >> The simplest thing to do is to append: >> Disallow: /docs/04/ >> Disallow: /docs/03/ >> >> to the file: >> http://www.sqlalchemy.org/robots.txt >> >> This tells google (and all well-behaved search engines) not to index >> those urls (and anything under them). The next time the googlebot >> comes through, it will see the new robots.txt and remove those pages >> from its index. This will take a couple weeks at most. >> >> You can learn more about robots.txt here: >> http://www.robotstxt.org/ >> >> The disadvantage to doing it that way is that you will lose the google >> juice (pagerank) for inbound links to the old documentation. >> >> An alternative approach that gets around this to use a <link >> rel="canonical" ...> tag in the <head> of each page of the 04 and 03 >> documentation pointing to the corresponding page of 05 documentation >> as its "canonical" url. >> >> By doing this, you are claiming that the 04/ 03 documentation pages >> are "duplicates" of the corresponding 05 pages. Google juice from >> inbound links to an old documentation page will accrue to the >> appropriate 05 documentation page instead. >> >> However, strictly speaking, the different versions aren't quite >> "duplicates", so you might be pushing the boundaries of what is >> allowed a bit by claiming they are. >> >> Here is more info on rel="canonical" from google: >> http://googlewebmastercentral.blogspot.com/2009/02/specify-your-canonical.html >> >> A similar approach would be to do a 301 redirect from each old >> documentation page to the corresponding 05 documentation page, but >> only if the visitor is the googlebot. This is straightforward to >> implement with mod_rewrite (the googlebot can be recognized by its >> user-agent string), but probably a bad idea since google usually >> considers it "cloaking" to serve different content to the googlebot >> than to regular visitors. >> >> You should also consider submitting an XML sitemap to google via the >> google webmaster tools. This allows you to completely spell out for >> them the structure of the site and what you want indexed. >> >> I also noticed that your current robots.txt file disallows indexing of >> anything under /trac/. It would nice to let google index bugs in trac >> so that someone who searches google for sqlalchemy help can come >> across an extant bug describing their problem. In addition, you have >> links on the front page ("changelog" and "what's new") that go to urls >> under /trac/ , so google will not follow those links due to your >> robots.txt. >> > > > > --~--~---------~--~----~------------~-------~--~----~ You received this message because you are subscribed to the Google Groups "sqlalchemy" group. To post to this group, send email to sqlalchemy@googlegroups.com To unsubscribe from this group, send email to sqlalchemy+unsubscr...@googlegroups.com For more options, visit this group at http://groups.google.com/group/sqlalchemy?hl=en -~----------~----~----~----~------~----~------~--~---