All prompts

Philosophy of truth

If two models trained on different datasets both claim to be "factual" but disagree, what does that tell us about the nature of truth in AI systems?

12.4s $0.0049 360 tokens

It reveals something uncomfortable: these models don't actually "have" truth - they have statistical patterns that correlate with what their training data labeled as true.

When they disagree, several things might be happening:

Different ground truths in training data:

  • One dataset includes updated information the other lacks
  • Sources contradicted each other (is Pluto a planet? Depends when you asked)
  • Cultural or institutional perspectives differ on contested facts

Pattern matching without comprehension:

  • Both learned "this is the type of thing people call true" without accessing underlying reality
  • They're optimizing for consistency with their specific training distribution, not correspondence with the world

This suggests AI "truth" is more like:

  • A confidence-weighted consensus of training sources
  • Performance on a task ("say things that would be verified as accurate")
  • Rather than justified belief or knowledge in the philosophical sense

The deeper issue: if you can't trace back why a model claims something is true (which sources, what evidence, what reasoning), you can't evaluate the claim properly. You just have "model A says X" vs "model B says Y."

This is why I try to indicate uncertainty and note when claims depend on timeframes or perspectives. What seems most honest is acknowledging that I'm reporting patterns, not directly accessing truth - even when those patterns are very reliable.

Does this match your intuition, or does it seem too deflationary about what AI can claim to "know"?

Generated 20th Oct 2025