Last modified: 2011-03-13 18:05:53 UTC

Wikimedia Bugzilla is closed!

Wikimedia has migrated from Bugzilla to Phabricator. Bug reports should be created and updated in Wikimedia Phabricator instead. Please create an account in Phabricator and add your Bugzilla email address to it.
Wikimedia Bugzilla is read-only. If you try to edit or create any bug report in Bugzilla you will be shown an intentional error message.
In order to access the Phabricator task corresponding to a Bugzilla report, just remove "static-" from its URL.
You could still run searches in Bugzilla or access your list of votes but bug reports will obviously not be up-to-date in Bugzilla.
Bug 13439 - Concealing search results that are listed in robots.txt
Concealing search results that are listed in robots.txt
Status: RESOLVED WONTFIX
Product: MediaWiki extensions
Classification: Unclassified
General/Unknown (Other open bugs)
unspecified
All All
: Lowest enhancement (vote)
: ---
Assigned To: Nobody - You can work on this!
:
Depends on:
Blocks: robots.txt
  Show dependency treegraph
 
Reported: 2008-03-19 22:54 UTC by Brian McNeil
Modified: 2011-03-13 18:05 UTC (History)
2 users (show)

See Also:
Web browser: ---
Mobile Platform: ---
Assignee Huggle Beta Tester: ---


Attachments

Description Brian McNeil 2008-03-19 22:54:04 UTC
The objective is to avoid having pages which are within a namespace that is searched by default and fall into one of the listed directories in robots.txt are, by default, suppressed in the returned results.

Ideally this would be optionally accessible to logged in users by giving a link at the head or foot of the results pages when there are hidden results, possibly "Show hidden results" followed by a "Why hidden?" link to explain what is going on.

This has been discussed on foundation-l and, for English Wikinews, would suppress the display of prepared obituaries (Eg go to http://en.wikinews.org and search for "Carter")
Comment 1 Brion Vibber 2008-03-19 23:05:10 UTC
Note that an existing way to handle that case would be to segregate such "prepared stories" in a namespace which is not searched by default.
Comment 2 Aryeh Gregor (not reading bugmail, please e-mail directly) 2008-03-19 23:11:52 UTC
This is not reasonably possible as stated.  The wiki software does not know whether a robots.txt exists or where it might be located, let alone what's in it.  I would be inclined to say WONTFIX on the basis that any fix for the problem as stated would be a horrible and fragile hack that could be much better implemented in other ways.  I would be doubly inclined to say so on the basis that robots.txt is not meant to hide things from everyone, only from robots, and robots don't read or index search pages anyway, so the relevance seems to be near-zero.

In particular, it seems like the best solution to your quandary is just to use a namespace that's not searched by default.

Note You need to log in before you can comment on or make changes to this bug.


Navigation
Links