Last modified: 2008-05-13 22:43:20 UTC

Wikimedia Bugzilla is closed!

Wikimedia has migrated from Bugzilla to Phabricator. Bug reports should be created and updated in Wikimedia Phabricator instead. Please create an account in Phabricator and add your Bugzilla email address to it.
Wikimedia Bugzilla is read-only. If you try to edit or create any bug report in Bugzilla you will be shown an intentional error message.
In order to access the Phabricator task corresponding to a Bugzilla report, just remove "static-" from its URL.
You could still run searches in Bugzilla or access your list of votes but bug reports will obviously not be up-to-date in Bugzilla.
Bug 13398 - Add bot generated spam reports on enwiki to robots.txt
Add bot generated spam reports on enwiki to robots.txt
Product: Wikimedia
Classification: Unclassified
General/Unknown (Other open bugs)
All All
: Normal enhancement (vote)
: ---
Assigned To: Nobody - You can work on this!
: shell
: 13529 (view as bug list)
Depends on:
  Show dependency treegraph
Reported: 2008-03-16 23:46 UTC by Alex Z.
Modified: 2008-05-13 22:43 UTC (History)
3 users (show)

See Also:
Web browser: ---
Mobile Platform: ---
Assignee Huggle Beta Tester: ---


Description Alex Z. 2008-03-16 23:46:15 UTC
There have been numerous complaints to OTRS about enwiki's bot generated spam reports[1] showing up high on search results, associating the site with spam, even though that isn't always the case, many things will result in a report being generated: account name is "similar" to a domain, IP "close" to the domain adds the link, and there's little differentiation between someone who adds one link and someone who adds 100. 

Adding the following to robots.txt in the section to disallow all user-agents should fix this:
Disallow: /wiki/Wikipedia:WikiProject_Spam/LinkReports/
Disallow: /wiki/Wikipedia%3AWikiProject_Spam/LinkReports/

Comment 1 Alex Z. 2008-03-27 01:55:24 UTC
*** Bug 13529 has been marked as a duplicate of this bug. ***
Comment 2 Ral315 2008-03-27 02:23:40 UTC
I would expand it to instead include the following:

Disallow: /wiki/Wikipedia:WikiProject_Spam/
Disallow: /wiki/Wikipedia%3AWikiProject_Spam/

That way, it includes everything under the main page, including a few pages that I think wouldn't be covered, but might need to be covered, by the robots.txt file.
Comment 3 Betacommand 2008-04-14 20:03:32 UTC
basic subpages should not be ignored the pages that cause the problems all start with Wikipedia:WikiProject Spam/Link so Wikipedia:WikiProject Spam/Link* and Wikipedia talk:WikiProject Spam/Link* should be added. ~
Comment 4 Brion Vibber 2008-05-13 22:43:20 UTC

Disallow: /wiki/Wikipedia:WikiProject_Spam/
Disallow: /wiki/Wikipedia%3AWikiProject_Spam/

Note You need to log in before you can comment on or make changes to this bug.