Last modified: 2008-09-13 00:48:23 UTC
Have received vague reports of nostalgia.wikipedia.org showing up unexpectedly in regular Google search results. (This holds a copy of Wikipedia's database from early 2002, displayed in the old-style 'Nostalgia' skin, and was put up for one of Wikipedia's anniversary celebrations a few years ago.) The nostalgia site appears to be served out of the primary document root, so gets the regular robots.txt; we should possibly give it a custom docroot with a blanked Disallow robots.txt, which would phase it out of general web search indexes.
We should just redirect robots.txt to extract2.php and have them edited via the web.
Hmmm, sounds kind of scary but would probably work fine. :)
Will robots follow redirects for robots.txt? I guess we should proxy the request to extract2.php.
Done. User-agent: * Disallow: /