Last modified: 2008-05-22 21:30:34 UTC

Wikimedia Bugzilla is closed!

Wikimedia has migrated from Bugzilla to Phabricator. Bug reports should be created and updated in Wikimedia Phabricator instead. Please create an account in Phabricator and add your Bugzilla email address to it.
Wikimedia Bugzilla is read-only. If you try to edit or create any bug report in Bugzilla you will be shown an intentional error message.
In order to access the Phabricator task corresponding to a Bugzilla report, just remove "static-" from its URL.
You could still run searches in Bugzilla or access your list of votes but bug reports will obviously not be up-to-date in Bugzilla.
Bug 12546 - robots.txt -- exclude ro:VfD
robots.txt -- exclude ro:VfD
Product: Wikimedia
Classification: Unclassified
General/Unknown (Other open bugs)
All All
: High normal with 2 votes (vote)
: ---
Assigned To: Nobody - You can work on this!
: shell
Depends on:
  Show dependency treegraph
Reported: 2008-01-08 13:08 UTC by Bogdan Stancescu
Modified: 2008-05-22 21:30 UTC (History)
5 users (show)

See Also:
Web browser: ---
Mobile Platform: ---
Assignee Huggle Beta Tester: ---


Description Bogdan Stancescu 2008-01-08 13:08:35 UTC
Please add the following to robots.txt, in order to exclude the votes for deletion on the Romanian Wikipedia:

# ro:
#<Bug #>
Disallow: /wiki/Wikipedia:Pagini_de_%C5%9Fters
Disallow: /wiki/Wikipedia%3APagini_de_%C5%9Fters
Disallow: /wiki/Discu%C5%A3ie_Wikipedia:Pagini_de_%C5%9Fters
Disallow: /wiki/Discu%C5%A3ie_Wikipedia%3APagini_de_%C5%9Fters
Comment 1 Tim Starling 2008-01-09 03:16:04 UTC
Is it only those four pages? If so, it's more easily done with $wgArticleRobotPolicies.
Comment 2 Bogdan Stancescu 2008-01-09 14:50:54 UTC
Actually it's only two pages, they're duplicated because I've seen other languages' entries with the column escaped and unescaped. Is there any way we can modify that variable locally?
Comment 3 JeLuF 2008-01-16 21:05:04 UTC

'wgArticleRobotPolicies' => array(
    'rowiki' => array(
    	'Wikipedia:Pagini de şters' => 'noindex,follow',
    	'Discuţie Wikipedia:Pagini de şters' => 'noindex,follow',
Comment 4 Bogdan Stancescu 2008-02-01 09:29:16 UTC
For some reason, the solution doesn't work. This bug was closed on January 16th, but I'm able to find a VfD page *created* on January 25th using Google:
Comment 5 Bogdan Stancescu 2008-02-10 00:52:07 UTC
Comment 6 Bogdan Stancescu 2008-02-19 11:40:40 UTC
Reminder. I changed the priority because we're getting pressure over this, I expect you know why this is being requested on various Wikipedia installations so I won't reiterate the rationale.
Comment 7 Razvan Socol 2008-02-22 08:32:42 UTC
I consider that this is not an "enhancement", but a "normal" bug that should be fixed, in order that behaves similar to most other wikis. We are really getting pressure over this (as it says in robots.txt: "Folks get annoyed when VfD discussions end up the number 1 google hit for their name. See bugzilla bug #4776"), so please add those lines to robots.txt as soon as possible.
Comment 8 Bogdan Stancescu 2008-04-19 20:52:21 UTC
It's been three and a half months since this bug was opened, and we can still find out VfD pages on Google:
Comment 9 Bogdan Stancescu 2008-04-19 20:58:04 UTC
Actually, that was an old VfD -- just checked with several newer ones and they don't show anymore. Closing bug.
Comment 10 Bogdan Stancescu 2008-05-17 23:24:58 UTC
Strangely enough, although on 2008-04-19 I was unable to find the VfDs I searched for, Google's behavior proves to be erratic: an old VfD which we had deleted on account of non-flattering comments on a living person re-surfaced on Google searches. Please, pretty please with sugar on top, will someone add those lines to robots.txt?
Comment 11 Brion Vibber 2008-05-19 16:23:35 UTC
I can confirm the <meta> robots tags on the *two and only two* requested pages *are* marked as noindex.

Did you really mean to request *those two and only two pages* or them *and all subpages*?
Comment 12 Razvan Socol 2008-05-20 09:55:56 UTC
I am sure that Bogdan means to request to add those pages *and all subpages*.
Comment 13 JeLuF 2008-05-20 18:07:32 UTC
Robots.txt does not affect pages that Google has already spidered. When a page that has already been deleted on the wiki shows up in Google, changes to our robots.txt will not change the search result.
Comment 14 Brion Vibber 2008-05-20 18:13:26 UTC
certainly implies that content will be removed from index once it's listed in robots.txt (after the site's crawled again, naturally).
Comment 15 Brion Vibber 2008-05-22 21:30:34 UTC
Updated robots.txt

Note You need to log in before you can comment on or make changes to this bug.