Last modified: 2006-10-21 16:38:31 UTC
LocalSettings.php would have a switch which would allow tiny bits of xml to be
stuck in the content of pages. ( $wgAllowRawXML=true||false; )
Example content formatting IN the articles might look like this:
<definition>A fantastic piece of software.</definition>
For allowing external applications (which could be web, or windows/mac/linux
client software) to not only pull full content out of MediaWiki (Similar to
functionality in API Bug 208 & XML export), but once they have the contents of
the article, they can pull detailed, specific data out of the article itself as
The above situation(s) could be universally useful to Wiktionary and Wikipedia.
Currently, a third-party company called Webaroo(.com) takes full database dumps
of Wikipedia to provide offline content to PDAs and laptops.
Imagine the power of being able to dynamically pull an entire dictionary, or
parts thereof, to your PDA (offline) and 'synchronize' it periodically? Surely,
these are only a few of the possibilites...
P.S. Is there already some sort of functionality/API/methodology that might
allow this, or something similar?
This can already be done in any number of ways, such as storing them in template
parameters that don't show up as anything. Templates by themselves are pretty
easy for anyone to parse; see also [[Wikipedia:Persondata]]. What advantage
would we get by disappearing anything unrecognized that looks like XML? And
where is this XML to be output, if anywhere?
You can do this trivially as a custom extension.
Of course this would never ever be allowed on a public site like Wikipedia since, being
identical to $wgRawHtml, it would be a huge security risk.