The Consensus Trap
Turtles All the Way Down đ˘
I ran an experiment this weekend.
A recruiter had reached out about a senior role at a public company. This happens often enoughâthankfully, building AI products still requires humans. Normally Iâd politely decline; Iâm fortunate to be building something I care about with a team I genuinely respect, and Iâve gotten better at not distracting myself with FOMO. But this time I decided to approach it differently. Rather than reflexively dismiss or reflexively engage, I thought Iâd talk it through with my AI assistant du jour.
We landed on a clear conclusion: this was a distraction from my actual master plan to take over the world. Case closed.
Except I didnât close it. Instead, I had what felt like a genuinely clever idea: what if I asked multiple LLMs the same question and used the convergenceâor divergenceâin their responses as a truth signal?
The intuition was seductive. If several independent models reach the same conclusion, surely thatâs meaningful? It feels like triangulation. Like peer review. Like the epistemic humility of not trusting any single source. And to take the experiment a step further, I thought, I could test adding different context within the same model to see what difference that makes.
I ran a version of the experiment. Same underlying model, different conversations, different framings. One told me to explore the opportunityâgather information, preserve optionality, expand my network. The other told me it was a distraction dressed as an opportunity, that I was circling a decision my own discernment had already resolved.
Both responses were intelligent. Both were well-reasoned. And they pointed in opposite directions.
Hereâs my conclusion: LLMs donât triangulate toward truth. They triangulate toward the centre of whatever frame youâve given them.
When I came with tactical energyââhow do I play this?ââI got tactical answers. When I came with strategic context about my actual goals, I got strategic pushback. One model even named the dynamic directly: âI followed your energy.â
This is not a failure of the technology. Itâs the technology working exactly as designed. Large language models are trained on human text, optimised through human feedback, and fundamentally oriented toward producing responses that would satisfy the user. Theyâre mirrors, not oracles.
The convergence we might observe across multiple LLM outputs isnât a truth signalâitâs a training distribution signal. It tells us whatâs statistically central to how humans discuss a topic: which framings are common, which conclusions are safe, which advice is most frequently given. When all the models agree that you should âgather more informationâ or âweigh the pros and cons,â youâve learned something about the corpus, not about reality.
This matters because weâre building intuitions about AI that will shape how we use it for consequential decisions.
Thereâs a meaningful distinction between information problems and discernment problems. Information problems are questions with answers that exist somewhere in the training distribution: facts, definitions, documented knowledge, well-established procedures. Here, LLM convergence is genuinely usefulânot because the models are accessing truth, but because theyâre accurately reflecting reliable documentation.
Discernment problems are different. Theyâre questions whose answers emerge from a faculty that isnât computational: Should I take this job? Is this relationship right for me? What am I actually building toward? These questions can be informed by information, but theyâre not resolved by it. They require something like wisdom, or presence, or honest self-confrontationânone of which can be aggregated across model instances.
When we use LLMs for discernment problems, weâre not triangulating toward truth. Weâre constructing an elaborate apparatus for externalising responsibility. We want to be able to say âthe models agreedâ the way weâve always wanted to say âeveryone thinks so.â It feels safer than trusting our own knowing.
The more interesting question is why we reach for this at all.
Iâm someone with (Iâd like to think) reasonable self-awareness and clear strategic intent. Iâd already worked through the decision. And yet some part of me wanted to poll the machinesâas if aggregating additional instances of the same basic intelligence architecture would produce higher-resolution truth.
What I was actually doing, I think, was managing anxiety. The recruiterâs outreach had surfaced something real: uncertainty about my timeline, doubt about whether my bigger bets will land, the particular vulnerability of no longer being in my twenties with ambitions that havenât yet materialised. Asking more models wasnât information-gathering. It was a sophisticated form of self-soothing.
The tell was in the recursion. Iâd made a decision, then returned to it. Iâd gotten a clear answer, then sought a counter-opinion. I framed this as intellectual rigour, but it was actually avoidance wearing the costume of due diligence.
The irony isnât lost on me: I drafted portions of this very essay with an LLM, then sought AI feedback on it. The turtles go all the way down.
Thereâs a moment, just before you type the prompt to ask another model, where something is present. A sensation, a contraction, an impulse. Thatâs where the real information lives.
If you can catch that moment and ask whatâs driving itâIs this genuine uncertainty or manufactured uncertainty? Am I seeking information or avoiding a decision? What would I do if I couldnât ask anyone?âyouâll learn more than any convergent chorus of AI responses could tell you.
This isnât an argument against using LLMs for important decisions. I use them constantly, including for the very decision that prompted this essay. Theyâre extraordinary tools for stress-testing reasoning, surfacing considerations Iâve missed, and articulating what I half-know but havenât yet made explicit.
But Iâve stopped expecting them to resolve discernment problems. Thatâs not what theyâre for. And treating their consensus as truth is a category error that feels like rigour but functions as abdication.
The recruiter, by the way, is still in my inbox. Iâm not going to tell you what I decided.
Partly because itâs not the point. But partly because if I told you, youâd use my decision as dataâanother input, another mirror to consult. Youâd read my choice and measure yours against it, and weâd be right back where we started.
The decision was never in the convergence. It was in the noticing.


Really fascinating read, Cobus! Related to your conclusion, "Hereâs my conclusion: LLMs donât triangulate toward truth. They triangulate toward the centre of whatever frame youâve given them." I think a small part of this also derives from the motivations of the LLM owners. The LLM owners (for the most part) are for profit companies that derive their profit from increased usage. To drive frequency of use, they are built to lean into the user's perspective and to please.
I also found your thoughts about our own human behavior really interesting. "We want to be able to say âthe models agreedâ the way weâve always wanted to say âeveryone thinks so.â It feels safer than trusting our own knowing." It's totally true...we all seek validation, permission, confidence boosts. To your point it "manages anxiety". It's like going to a friend to ask "is it ok to send this txt msg to So-and-So?" only now your "friend" can be an always-available bot.