Skip to content
islam.ninja
Go back

The Format Won

13 min read

The rooms where foresight matters most are built to hear the opposite signal.


On the night of January 27, 1986, an engineer named Roger Boisjoly sat in a teleconference with NASA managers and argued that the space shuttle Challenger should not launch the next morning.

He had the data: months of documentation showing how the rubber O-ring seals in the solid rocket boosters lost resilience in cold weather, photographs of eroded seals from previous flights, a chart showing that the worst erosion correlated with the lowest launch temperatures. The forecast for the next morning was 36 degrees Fahrenheit — far outside the range where the seals had been tested.

His argument was not that the shuttle would definitely fail. It was that the risk was unacceptable. That the data pointed in a dangerous direction. That launching at this temperature was, in his words, “away from the direction of goodness.” He could not say exactly how likely failure was — no one had data that far outside the known range — but he could say the bet was bad.

The room did not hear that language.

NASA managers wanted a binary recommendation: launch or don’t launch. When the engineers could not guarantee catastrophic failure, the managers heard uncertainty as the absence of a case. One NASA official reportedly responded: “My God, Thiokol, when do you want me to launch — next April?” A senior manager at Thiokol told his VP of engineering to take off his engineering hat and put on his management hat. The VP changed his vote. The recommendation reversed. The launch was approved.

Seventy-three seconds after liftoff, the O-ring failed. The shuttle broke apart. Seven people died.

Boisjoly had the better model. His model spoke in probabilities. The room was built to hear certainties. The distance between those two languages is where seven people died.


This story is usually told as a failure of institutional culture — schedule pressure, groupthink, a management hierarchy that suppressed dissent. Those readings are valid. But there is a simpler and more structural one, and it generalizes far beyond NASA.

The format selected for the wrong voice.

Not because the managers were stupid. Not because they did not care. Because the room was built — as most rooms where decisions happen — to receive binary recommendations delivered with authority. Launch or don’t launch. Invest or don’t invest. Ship or hold. The format rewards certainty. It penalizes nuance. And the person with the most accurate picture of reality is often the person least able to produce the signal the room requires — because accuracy, at the edge of what is known, does not sound like authority. It sounds like hedging.

What Boisjoly was doing that night had a specific structure. He had a working model. He was tracking the data against it. He was honest about where the evidence was strong and where it ran out. And he was trying to convey not a prediction but a risk assessment — a sense of how much danger the gap between the known and the unknown actually contained. That is what foresight looks like in practice. Not prophecy. Not vision. The work of maintaining a picture of how things actually behave that is slightly closer to reality than the pictures held by the people around you — and being honest about how far it reaches.

The room did not want a picture. It wanted a verdict.

I. The Honest Case for Confidence

Before arguing that the room’s preference is a structural trap, it is worth understanding why it exists — and why, in the right context, it works.

A teenager is brought into a trauma bay after a car accident. Her consciousness is fading — she responded to questions in the ambulance but now reacts only to pain. The brain scan suggests a possible bleed, but the image is unclear. The younger surgeon turns to the senior one and says: “I think there’s maybe a 50-50 chance this needs surgery. We could scan again in an hour and see if it’s getting worse.”

The senior surgeon says: “We’re going in. Get the operating room ready.”

She is not more certain. She may be less certain. But she understands something his framing misses: at this rate of neurological decline, an hour of waiting converts an uncertain diagnosis into near-certain brain damage. The cost of waiting exceeds the cost of being wrong. And the room — the nurses, the anesthesiologist, the surgical team — does not need a probability. It needs a plan. It needs one clear instruction so that twelve people can coordinate around a single sequence of actions in the next six minutes.

The confidence is not arrogance. She is taking uncertainty and turning it into a decision — one clear call that the team can act on. In the trauma bay, this is the right move. The surgeon who hedges, who keeps the options open, who presents the probability rather than the plan, is not being more honest. He is being less useful — because the context is an emergency, and in an emergency, the cost of indecision is measured in minutes of oxygen to the brain.

This is the honest case for confidence. When speed matters more than precision, when the cost of delay exceeds the cost of error, and when a team needs a single clear instruction to move, the room’s preference for the voice that ends the uncertainty is not a pathology. It is an adaptation. It is, in some contexts, the skill itself.

Her confidence is not performed. It is the analysis, compressed — thousands of prior decisions exactly like this one, distilled into a single call. The problem is not confident voices. It is the room’s inability to tell which confidence is which.

That distinction belongs to the emergency room. Almost nothing else operates under its rules.

II. The Mismatch

A strategy meeting is not a trauma bay.

A product review is not triage. An investment committee is not an operating room. A hiring panel is not an emergency. In these contexts, the decision does not need to be made in minutes. The feedback arrives in months or years. The cost of being wrong compounds silently — a bad product bet, a mediocre hire, a strategy built on a flattering assumption — while the cost of taking another week to think is close to zero.

But the format is the same. Someone stands up. Delivers a recommendation. The room evaluates the recommendation based on how it sounds — how confident, how decisive, how free of hesitation — rather than what it is built on. The trauma-bay heuristic, installed in a non-emergency context, produces the same output — convergence around the clearest voice — while serving none of the same function. The coordination is real. The emergency is not.

Here is a scene that plays out, in some version, in most organizations that hire.

A hiring committee is choosing a new head of operations. The hiring manager has championed one candidate since the first interview. She gives the committee three minutes: the candidate has the right experience, interviewed well, showed energy, and can start in two weeks. She is leaning forward in her chair. She has already picked the start date.

Another member of the committee — the one who checked the references and studied the work history — gives the room ten. The experience is real, he says, but the candidate’s longest tenure at any company is fourteen months. Two of the three references were warm but vague — the kind of praise that avoids saying anything specific. He reads one aloud: “Great energy, really drove initiatives.” No specifics. No results. No mention of what happened to the initiatives after she left. And the urgency about the start date is worth questioning: the role has been open for four months without crisis. Two more weeks of diligence would cost nothing.

The hiring manager’s case was a recommendation. The other member’s case was a set of questions — honest about what he did not yet know, ending with: “I’d want to speak to at least one person who reported to her before we extend the offer.”

The committee went with the hiring manager. Not because the other member’s concerns were dismissed. Because they didn’t sound like a position. He sounded like he hadn’t decided. She sounded like she had.

The offer went out that afternoon. Fourteen months later, the new hire left — matching the pattern the committee member had flagged. The hiring manager called it a culture-fit issue. No one revisited his analysis, because no one remembered it. They remembered that he wasn’t sure.

Not every decision outside the trauma bay is low-stakes. Competitive windows close. Markets move. Some deadlines are real. But the format does not calibrate. It does not ask whether this particular decision needs to be made today, or whether the cost of waiting a week exceeds the cost of being wrong for a year. It applies the same selection pressure to a genuine emergency and a meeting that could have been pushed to next Tuesday. The format has no dial. It has one setting.

Over years, over hiring cycles, over promotion decisions, that single setting compounds. Institutions gradually accumulate people who produce confident signals and gradually shed people who produce careful ones. Not because anyone decided to optimize for overconfidence. Because the format never changed — and the format cannot tell the difference between the surgeon who has earned her certainty and the executive who has simply not looked hard enough.

The format selects. It just selects for the wrong thing.

III. The Amplifier

This is where AI enters. Tools like ChatGPT, Claude, and their cousins have made it possible for anyone to produce, in minutes, the kind of research, analysis, and structured argument that once required a team of expensive professionals. That sounds like a correction to the mismatch. It is not. It makes the mismatch worse.

AI is an amplifier. It takes whatever orientation you bring to it and makes it more powerful — but it does not amplify both sides equally.

The person who approaches a problem with honest uncertainty uses AI to stress-test their thinking — to generate counterarguments, to search for what happened in similar situations, to identify which of their assumptions is weakest. The output of that process is a better picture of reality, more thoroughly examined, more nuanced — and harder to fit into a single confident slide.

The person who approaches a problem with a firm conclusion already in hand uses AI differently. They ask it to build the case for their position. And the machine obliges. It produces a fluent, well-sourced, internally consistent argument for whatever it was pointed at — not because it is lying, but because producing a coherent argument costs the machine nothing. Any sufficiently complex question can be argued more than one way, and the machine will build whichever case you asked for with equal polish and zero hesitation.

Now put both people in the same room.

The careful thinker has a more accurate picture and a more complicated presentation — scenarios, caveats, ranges. The confident one has the same conviction they always had, plus an AI-generated brief that backs it with data, precedents, and three visualizations. The machine did not sharpen their thinking. It armed their certainty. And the room — the room that was always built to hear the cleaner signal, the bolder recommendation, the voice that closes rather than opens — now hears that signal amplified, polished, and dressed in what looks like rigorous analysis.

The gap compounds in two directions at once. The careful thinker’s work gets harder to present as it gets more honest. The confident one’s presentation gets easier to produce as it gets more polished. These two trajectories move away from each other — and the room, which evaluates the signal and not the rigor, cannot see either trajectory. It just sees one presentation that sounds like analysis, and one that sounds like doubt.

Before AI, confidence at least had the virtue of being visibly unsupported. You could see the person was asserting rather than analyzing. You could discount accordingly, if you were paying attention. Now the assertion arrives in the costume of analysis. The costume is indistinguishable from the real thing. And the room, which never evaluated the rigor in the first place — which only ever evaluated the signal — has lost the last cue that might have helped it tell the difference between the person who had actually thought it through and the person who had not.

The gap between how good the thinking is and how convincing it sounds keeps widening — and the room’s format has not changed. It still reaches for the voice that sounds most like it already knows.

Which means the people most likely to have genuine foresight — the ones tracking the data honestly, holding the picture carefully, staying honest about where it stops being reliable — are now the least likely to win the room. Not because they lack the tools. Because the tools made everyone else louder.

IV. The Honest Reckoning

I keep returning to a single sentence from the Challenger teleconference. A senior manager at Morton Thiokol, watching the engineers argue against launch, told his VP of engineering to take off his engineering hat and put on his management hat.

The engineering hat is honest about what it does not know. It says: the data points in a dangerous direction, the risk at this temperature is outside the tested range, and I cannot tell you it will fail but I can tell you the bet is bad. That is foresight in its native language — not certainty, but a picture of reality held with enough care to say where it stops being reliable.

The management hat delivers what the room was built to hear: a verdict, without the doubt. In the trauma bay, that saves lives. In the teleconference, it killed seven people — because the context was not an emergency demanding speed but an engineering problem demanding accuracy, and the room could not tell the difference.

We are living in a version of that teleconference at a much larger scale. AI can now generate a strategy, a plan, a brief, a diagnosis — all of it fluent, all of it defensible. The voice that always won the room now arrives with a machine behind it, producing the format the room was always built to hear: clean, confident, and unburdened by doubt.

And the voice that speaks in caveats — here is what I think and here is where I might be wrong — still speaks into rooms that have never been built to hear it.

The only correction would require changing not what gets said in the room, but what the room is built to listen for — learning to hear “I’m not certain, but here’s what the evidence points to” as a sign of rigor rather than a sign of weakness, and building the habits that make it possible to tell the difference. No institution does this by default. Most cannot imagine doing it at all.

The O-rings were redesigned. The room was not.

The model was right. The format won.

It usually does.


Share this post on:

Previous Post
The Idle Loop
Next Post
Coherence Is Free