Signals & Subtractions #033: The Copilot Problem

Signals & Subtractions #033: The Copilot Problem

Jan 12, 2026 | Issue 33

tbd


Issue-033


🔭 Signal: When Politeness Becomes a Hazard

A recent story about a physician and an AI tool has sparked debate about whether AI belongs in clinical settings. But the incident pattern it describes isn’t new, and it isn’t about AI.

A high-status professional acts alone. An informal peer-check gets bypassed. A single source substitutes for consultation. Validation happens late or not at all. The error looks “reasonable” in context.

Calling this “misuse” is insufficient, and calling it “system failure” without naming roles is also incomplete.

What broke here wasn’t judgment. It was structure. An unverified answer was allowed to bypass the validation loops that medicine already depends on everywhere else. The tool may be new, but the failure mode is ancient.

AI does make this worse, certainly. But in this case it’s not because AI is uniquely dangerous. All it does here is collapse the buffer that once allowed this kind of structural gap to remain survivable. When tools are fast, fluent, and embedded in real workflows, delayed validation changes outcomes.

The question isn’t whether AI should be trusted. It’s why the system allowed a single source (any single source) to stand alone in a safety-critical decision.

Nothing here is exotic. And that’s precisely the problem.


🧠 Strategic (Human) Prompt: Completion Criteria

Instead of asking “who made a mistake?” ask:

Did the system allow a single source to be sufficient?

If yes, the vulnerability existed before anyone touched a tool. The tool just found it faster.

This reframe matters because it shifts attention from blame to design. The practitioner isn’t the failure point. The missing structure is.


➖ Subtraction: Default Trust

This week, consider subtracting one or more of the following:

  • The assumption that a single “human in the loop” prevents single-source failures
  • The belief that training changes behavior when time pressure is high
  • The habit of treating politeness and deference as neutral
  • The comfort of post-hoc accountability instead of structural prevention

If a high-stakes task can be completed alone, quickly, without contradiction, then eventually it will be. Not because practitioners are reckless, because systems that reward throughput will find every gap in validation structure.

These assumptions feel protective, but in fact they’re dangerous.


✈️ Analogy of the Week: The Copilot Who Wouldn’t Speak

Tenerife, Canary Islands, 1977. Fog. Two 747s, same runway. The KLM captain begins takeoff without clearance. His flight engineer questions it, but…tentatively. The captain dismisses him.

583 dead. Still the deadliest accident in aviation history, nearly 50 years later.

The inquiry didn’t find a reckless pilot. It found a system where junior crew were trained to defer, and authority was allowed to complete actions unchallenged.

Aviation didn’t fix this by telling copilots to be braver. They redesigned authority itself. Crew Resource Management made challenge structural. Checklists require two voices. Callouts require responses. Ever since then, silence is treated as a system failure, not a personality flaw.

The aviation rule change didn’t make people any more courageous. They made courage unnecessary by making silence structurally impossible.

That’s the design question AI is forcing on every other industry now.


♬ Closing Notes

AI didn’t create single-point-of-failure workflows. It just removed the human slack that made them survivable.

Trainings, warnings, professional norms, and post-hoc accountability only work when:

  • time pressure is low
  • tools are slow
  • authority is distributed
  • errors surface early

AI breaks all those assumptions.

The debate over whether AI “belongs” in high-stakes settings misses the point. The structural gaps were already there. AI just operates fast enough to find them before humans can intervene.

The fix isn’t better training or louder warnings. It’s designing systems where no single source is ever sufficient, authority is intentionally incomplete, and validation is structural, not optional.

Until next week,

Sam Rogers
Structural Safety Advocate
Snap Synapse – from AI promise to AI practice

📅 Book a meeting
🎯 Measuring collaboration behaviors, not just individual capability: PAICE.work


When Politeness Becomes a Safety Hazard

Signal: Politeness Can Be Hazardous

AI collapses the buffer that once allowed unsafe judgment to remain survivable.

For decades, safety culture relied on a quiet compromise:

  • speak gently in public
  • diagnose bluntly in private
  • fix systems later

That worked when mistakes moved slowly. AI removes that slack.

When tools are fast, fluent, and embedded in real workflows, delayed truth-telling is no longer neutral. It changes outcomes.


The Incident Pattern

A recent WSJ article (de-gated here) about the pressures & perils of AI use for doctors makes points that hit home for all of us. But this isn’t about one tool or one profession.

The recurring pattern looks like this:

  • a high-status actor acts alone
  • an informal peer-check norm is bypassed
  • a tool substitutes for consultation
  • validation happens late or not at all
  • the error looks “reasonable” in context

Nothing here is exotic. That’s precisely the problem.


What Actually Failed

Not intelligence. Not intent. Not ethics.

Three things failed:

  1. Plurality = A single source was allowed to matter.
  2. Role boundaries = A professional acted outside the social checks that normally constrain them.
  3. Structure under pressure = Efficiency made skipping validation feel rational.

Calling this “misuse” is insufficient. Calling it “system failure” without naming roles is incomplete.


Why This Keeps Happening

Most systems still rely on:

  • training
  • warnings
  • professional norms
  • post-hoc accountability

Those only work when:

  • time pressure is low
  • tools are slow
  • authority is distributed
  • errors surface early

AI breaks those assumptions.


The Design Mistake We Keep Repeating

We keep designing for good behavior instead of bounded authority.

If a task can be completed:

  • alone
  • quickly
  • without contradiction
  • without peer involvement

then it eventually will be.

Not because people are reckless. Because systems reward throughput, not reflection.


What Safer Systems Do Differently

Across medicine, aviation, and other high-reliability fields, safer systems:

  • enforce plurality instead of encouraging it
  • require independent confirmation
  • prevent single-source completion
  • treat hierarchy as a hazard surface
  • make unsafe shortcuts noisy, slow, or incomplete

None of this relies on trust. It relies on constraint.


Subtractions

If you’re working with AI in safety-critical contexts, consider subtracting:

  • the assumption that “human in the loop” is sufficient
  • the belief that training changes behavior under pressure
  • the idea that politeness is always neutral
  • the habit of treating role violations as edge cases
  • the comfort of private correction after public silence

These feel reasonable. They no longer scale.


A Practical Question to Carry Forward

Before accepting an AI-assisted output, ask:

Could this decision have completed without anyone else involved?

If the answer is yes, the system is already unsafe. Not because of AI. Because singularity slipped in.


Closing

This isn’t about banning tools or blaming people.

It’s about designing systems where:

  • no one is sufficient alone
  • authority is intentionally incomplete
  • validation is structural, not optional

That’s how safer judgment is practiced. At speed. Under pressure. By humans.

Read on Substack

Get in touch

Book a call, send a message, or connect on LinkedIn.

Book a meeting Contact us LinkedIn