h a l f b a k e r yQuis custodiet the custard?
add, search, annotate, link, view, overview, recent, by name, random
news, help, about, links, report a problem
browse anonymously,
or get an account
and write.
register,
|
|
|
WikiTrue
A wiki all of whose content is necessarily true | |
Wikipedia is a flawed but useful resource. However, when i look at it i realise there are several logical categories of content on it. There is content which is generally regarded as true, content which is popularly regarded as true, controversial content and content whose truth value is probably not
very important. Then, there's a category with a much higher standard of truth, which could be wrong of course, but whose wrongness can be clearly demonstrated using only the data on the site, for instance it might contradict itself or say different things in different entries, so they can't both be true. This would include articles on mathematics and logic, arguments, proofs and an entirely different category concerning cultural artifacts, for example Buffy is necessarily a fictional character in 'Buffy The Vampire Slayer' in a sense, i.e. there is canon and non-canon.
So my idea is this. As well as having a site like Wikipedia, whose content can be contentious and depends on dirty, muddy, confusing physical reality for a lot of its articles, over which users can argue endlessly, produce a website whose policy is only to present necessary truths - facts which cannot be rationally doubted and are true in all possible worlds. This could still contain a lot of Wikipedia-like content, including articles on calculus, fictional entities and phenomena which are however consistent within that fictional universe, and that which is true given the context of contentious issues. It could include articles on mythology, unicorns, Star Trek and the like as well. It could even include articles about things which are necessarily false, maybe with the introduction "The following article is false" in order to make it true.
Clearly people could and would argue about the necessary truth or otherwise of the content, leading to deletions but also, software could parse the text in such a way that any statement or conjunction of statements in an article would automatically be deleted if its truth table wasn't entirely consistent, or in fact couldn't even be submitted.
The value of this would be that it would be easier to trust the information in this Wiki without ploughing one's way through the whole lot to verify it. The scope would of course be limited, but the standard of reliability would be much higher.
Real Time Google Earth
http://www.popsci.c...-people-cars-clouds [JesusHChrist, Mar 12 2011]
Does "consistent with canon" imply "internally consistent?
http://books.google...0Musketeers&f=false [mouseposture, Mar 13 2011]
W3C: Semantic Web
https://www.w3.org/...ds/semanticweb/data [zen_tom, Jul 06 2016]
W3C: OWL Web Ontology Language
https://www.w3.org/...semantics-20040210/ This is the specification by which you can define your logical constructs that can be applied to information to measure consistency. [zen_tom, Jul 06 2016]
http://wiki.dbpedia.org/
http://wiki.dbpedia...about/facts-figures The English version of the DBpedia knowledge base currently describes 4.58 million things, out of which 4.22 million are classified in a consistent ontology (http://wiki.dbpedia.org/Ontology2014), including 1,445,000 persons, 735,000 places (including 478,000 populated places), 411,000 creative works (including 123,000 music albums, 87,000 films and 19,000 video games), 241,000 organizations (including 58,000 companies and 49,000 educational institutions), 251,000 species and 6,000 diseases. [zen_tom, Jul 06 2016]
Please log in.
If you're not logged in,
you can see what this page
looks like, but you will
not be able to add anything.
Destination URL.
E.g., https://www.coffee.com/
Description (displayed with the short name and URL.)
|
|
In context perhaps. No quarrel with internal consistency even if the entire system turns out to be false. |
|
|
Why would it be easier to trust the articles on this Wiki more than the same articles on any other Wiki? |
|
|
// Why would it be easier to trust the articles on this Wiki more than the same articles on any other Wiki? // |
|
|
Because it would be impossible to compose and submit an article which wasn't true. For instance, suppose you wanted to submit the text "two plus three equals one". The software running the site would turn that into a bit of arithmetic, note that it was false and reject it. If there was already an article consisting of the text "two plus three equals five" and you edited the last word to read "six", you would be unable to submit that edit. Anything you did to make the article inconsistent would also be rejected provided the software could detect that it was, so for example, there is an article containing the sentence "all unicorns always have blue eyes", to which you want to add the comment somewhere in the article, or perhaps anywhere in the wiki, "all unicorns always have brown eyes and never have blue eyes" or words which imply that logically. That edit would be rejected. |
|
|
WikiConsistent, perhaps. Also, mfd truth algorithm
magic? |
|
|
Do I understand correctly that, if I post "unicorns always
have blue eyes" then your subsequent attempt to post
"some unicorns have brown eyes" will fail? But that if we
had posted in opposite temporal order, mine would have
been the rejected post? |
|
|
Assuming this idea were even possible, why not vary the
rules by which edits are rejected, and use it as a testbed
for epistemological theories? For example, suppose it
contained a theory, and some observations mostly
consistent with the theory. Suppose more observations
are posted, not so consistent with the theory. Epicycles
are added to the theory, keeping it consistent with new
observations. Could you implement a system such that,
eventually someone posts a theory contradicting the old
one, but so much more consistent with the accumulated
observations that the new theory is accepted and the *old*
one rejected? |
|
|
Yes, you do understand me correctly. However, that would only be one genre of article on the wiki. They would simply be castles in the air, as it were. They wouldn't need to be true, only consistent. However, if there were entirely non-S-intensional subjects such as mathematics on there, they would have to be true because it would be impossible to post contingent truths on them at all. It would be theoretically possible to "parse" the whole lot without any empirically verifiable information and therefore also possible to rule out the submission of any false statements whatsoever in such subjects. |
|
|
I suppose there are at least two categories here: the necessarily true and the internally consistent. The necessarily true can be handled. That's not magic. It simply needs the truth value of each statement and the conjunction of every statement in an article to be true under all truth conditions of the components of the statements. |
|
|
[Mouseposture], yes, i like that idea. In fact, i suppose what i might have done here is mixed up two types of content, which i nonetheless feel sort of belong together. The thing is, when i look at Wikipedia, i often doubt its validity, but there are at least two types of article where that doesn't apply in the same way. One is mathematical articles where it's possible to check their truth out for themselves, for instance i haven't looked but i can imagine an article on prime numbers which would be necessarily true and checkable simply using computing power or even your own brain power. The other is more contentious, but consists of the likes of popular culture, myths and the like, where the settings and characters do not exist or are fictionalised, but cannot behave logically inconsistently without violating canon. |
|
|
Your idea is akin to the creation of a detailed fictional setting, but more rigorous, and it's really good! |
|
|
//cannot behave logically inconsistently without violating
canon// Not so fast. |
|
|
Have you read Umberto Eco's essay on this topic? <link> (I
recommend it: it's light and amusing, which is remarkable,
considering the subject.) |
|
|
//Because it would be impossible to compose and submit an article which wasn't true.// |
|
|
So there wouldn't be much content on this Wiki, it seems. |
|
|
I was reading this page, thinking about annotating
or voting on this idea, and just then my phone
chimed and I got my thank-you e-mail from
Wikipedia for my monthly donation. |
|
|
On topic, for this wiki to work, it seems like it
would need a branching category tree, where the
validity of an article is only checked against its
parent article, grand-parent, etc. Otherwise you
can't have, say, an article on Buffy the Vampire
Slayer and an article on Twilight, because vampires
are defined differently in those fictional worlds. |
|
|
So instead of having a wiki where you know
everything is true, you'll always know in which
context something is true by following the lineage
of the article. If it's attached straight to the root
or orphaned, you can probably disregard it. If it
has a long and venerable lineage of articles, it's
probably solid. |
|
|
Yes, i like that. Probably could be covered by a graphical representation of in which world the article was true. |
|
|
[Ldischler]. on the contrary. It's feasible to write entire books which can be mathematically proven to be true. Principia Mathematica (Russell and Whitehead, not Newton) appears to be one of these. I suspect there are articles on Wikipedia which have quite large necessarily true blocks of text in them. |
|
|
I was going to link to wikitruth.org, as it's kind of relavent and shows some of the pitfalls in the idea of "truth" in this context. |
|
|
The wikitruth website is gone. The "wikitruth" article on wikipedia has been deleted, along with its entire revision history and discussion page. Pitfalls indeed! |
|
|
The original dialogue basically went:
Wikitruth: Wikipedia is false.
Wikipedia: Wikitruth is false.,
which was bound to cause problems. |
|
|
I'm almost afraid to make this annotation, lest Jimmy's wrath descend on the Halfbakery in a Gödelian firestorm of paradoxical fury. |
|
|
It's almost the paradox of the preface: the conjunction of all statements made on Wikipedia is false but probably the majority of users believe the edits they make on Wikipedia result in statements whose conjunction is true. |
|
|
I love it!
But I think it needs to have a proof for each article,
and if necessary, a list of axioms that the encyclopedia accepts.
The proofs can be checked and edited by the editors for accuracy.
Maybe there should be a draft status for articles that are probably true but have errors in their proofs. This is to allow their authors to repair them, and it should be clear that they are not asserted by the wiki to be true.
The axiom list can be set to the common basis that people use for mathematics, logic, and possibly science - preferably intuitive stuff that laymen and mathematicians can agree on, rather than something like using ZFC and modelling numbers as sets of sets.
But at the same time, the axiom list needs to be carefully managed by the editors - the quality of their wiki depends on it. |
|
|
Secondly, mouseposture's idea of a wiki "testbed for epistemological theories" should also exist. It could be a sister project to WikiTrue, sharing editors and using WikiTrue articles as references where applicable.
In fact, I can't understate how good an idea I think the theory wiki is. It could start by dumping loads of commonly accepted scientific findings onto itself, but with each one treated as an unfounded hypothesis until the observations start coming in. |
|
|
Idischler says "So there wouldn't be much content on this Wiki, it seems."
That is true, but everything on it will carry a heavy weight of reliability! |
|
|
Even if the project is impossible because of Gödel's Incompleteness Theorem, it will be great, because we'll all get to see the effects of that theorem in action. |
|
|
The truth is you'll have to work it out for yourself. |
|
|
I'm currently working on a project that does exactly this! |
|
|
In addition to the WWW, Tim Berners-Lee also came up
with the idea of "Linked Data" and/or "Semantic Web"
which aims to provide a framework not of a linked web of
documents (as we all know and love), but a
linked web of data points. |
|
|
It starts with a specification for
describing knowledge, called Resource Description
Framework (RDF). RDF breaks down any element of data
into a "triple". A triple is an atomic unit of information,
and consists of a subject, a predicate, and an object.
Examples might be:
Luke isA Person
R2D2 isA Droid
Vader isA Person
Tatooine isA Planet
Luke isChildOf Anakin
|
|
|
Each element of the triple is referenced by URI. So in
actual fact, the first triple expressed in RDF might look
like: |
|
|
http://gfa.ents/Luke_Skywalker
http://gfa.preds/basic/isA
http://gfa.ents/Class/Person
|
|
|
Using URIs like this means that different people can generate data
about the same thing, from different locations, but can still cross
reference their federated information. The tricky part is agreeing
on standards - but just like WWW URLs, it's possible to
dereference, apply routing and aliasing etc where necessary to
map from one space to another. |
|
|
In addition to these factual statements, you can also
define something called an Ontology, itself defined in RDF
terms, in a language called Web Ontology Language (OWL) that, in addition to
describing the types of relationships you'd expect to see a particular entity have in
a particular context, also allows you to describe logical constraints and
connections between objects such as: |
|
|
isChildOf applicableTo Person
isChildOf NotApplicableTo Droid
isChildOf owl:inverseOf isParentOf
|
|
|
Not only can application of an ontology help identify
inconsistent additions to the data such as: |
|
|
Which would be identified as incorrect, or at the very
least, inconsistent when held up against the combination of this data-set and this particular ontology. |
|
|
Also interestingly is that using the ontology, and applying
it to a dataset, you can generate a whole new set of
triples based on the logical inferences on the intersection
between fact and logic. |
|
|
So with the examples, the ontology above and some
additional information like: |
|
|
You can generate the inference that:
|
|
|
Which means you can not only classify truthiness or
intruthiness, but you can also discover new truths for
which no explicit data already existed! |
|
|
I'll post some links, but it is very much a thing, and a
fascinating one to boot. |
|
|
That's just excellent! Thanks! |
|
|
You're very welcome - the standard for querying such data-stores is called SPARQL (pronounced Sparkle) - other options include Tinkerpop, Gremlin and Giraph. |
|
|
This can be quite entertaining as uninitiated members from outside the team are exposed to all these seemingly whimsical terms in the context of apparently hard-nosed delivery conversations. |
|
|
Can I just point out that a Web Ontology Language
would be a WOL, not an OWL. |
|
|
[zen_tom], what's gfa? And your description reminds me
strongly of Evi (formerly True Knowledge), which, back
when it was called True Knowledge, did that in a way
that anyone could contribute entities and facts to. |
|
|
I would also like to point out: |
|
|
- Microformats: A way to make data on normal web pages
semantic |
|
|
- Semantic MediaWiki: A set of extensions to MediaWiki
(the software that runs Wikipedia and loads of other
wikis) that make the wiki's content database-driven and
computable |
|
|
[notexactly], gfa is the www equivalent they use all the way over there in a Galaxy Far (Far) Away, it's usage is now obsolete (on account of it having been a long time ago) |
|
|
The relevance is less as something that gets embedded within in html documents (which is how many of these ideas were first couched), and more to do with sourcing data from open data portals. |
|
|
If you're a large institution and you're fed up of writing reports that everyone ends up hating, a nice way of sorting out the mess is to open up your data via a SPARQL endpoint, and let people query the hell out of it in their own time. |
|
|
Using a standard specification like RDF means less time fiddling about with defining schemas and upsetting people when you decide to change your attribute naming convention from underscore_spaces to CamelCase - RDF doesn't care because they've already specified that you use the most horrendous naming convention ever, the URI. |
|
|
"Hitler was evil" aaaaaand go! |
|
| |