The Assumption Bias Mitigation Protocol: A Leader's Framework for Verifying AI
- Severin Sorensen

- Oct 22
- 6 min read
Companion article to: "The AI Confidence Trap: When 85% Certainty Is Dangerously Wrong"
Your AI will deliver a sophisticated analysis with 85% confidence. You will act on it. And the recommendation may be catastrophically wrong.
This happens because AI confidence measures pattern matching, not information completeness. High confidence paired with low context is a recipe for systemic, high-stakes errors.
The solution is not to discard these powerful tools, but to impose discipline upon them. You must train your AI to pause, to question, and to verify before it recommends action.
This article provides the operational framework to do so. The Assumption Bias Mitigation Protocol is a set of principles designed to be embedded directly into your AI workflows. It translates the human disciplines of critical thinking and scientific inquiry into instructions the AI can understand and execute, protecting your organization from the dangers of false confidence.
The 7 Principles of the Mitigation Protocol
This protocol works by forcing the AI to deconstruct its own reasoning and reveal its own blind spots before presenting a final recommendation.
1. Separate Confidence from Completeness (The 40-Point Rule)
The protocol’s first rule breaks the illusion of certainty. It mandates that the AI explicitly state two different metrics:
Pattern Confidence: "I am X% confident this situation matches pattern Y."
Information Completeness: "I have Z% of the information I ideally need to act on this."
This creates the 40-Point Rule: If the Gap (Confidence % - Completeness %) exceeds 40 points, the AI is prohibited from issuing a recommendation. Instead, it must stop and generate questions to close the information gap.
2. Mandate Questions Before Conclusions
When confidence is high but completeness is low, the AI must automatically generate 3-5 critical questions. These are not simple clarifications; they are designed to falsify the initial hypothesis. The AI must be trained to skip to recommendations when the gap exceeds 40 points. Required questions include:
What information am I missing that would change this assessment?
What's the simplest explanation I'm overlooking?
What's the base rate for this outcome in similar situations?
What would prove this interpretation wrong?
If I'm wrong, what are the consequences?
3. Require the AI to Deconstruct Its Reasoning
To prevent "black box" thinking, the protocol requires the AI to clearly separate four distinct levels of analysis:
What I observed:
Objective data only: "Sales dropped 40%."
What I'm inferring:
Interpretation: "Productivity has declined."
What I'm assuming:
Gaps being filled: "This indicates disengagement."
What I don't know:
Recognized gaps: "I do not know their personal circumstances, baseline work patterns, or peer feedback."
4. Insist on a Base Rate Check
Left to its own devices, an AI will over-index on the specific case presented. The protocol forces it to anchor its analysis in statistical reality by stating the base rate.
Reference Class:
"This situation belongs to the category of 'top sales reps with sudden 40% performance drops.'"
Base Rate:
"In this reference class, 60-70% of cases are due to temporary external factors (e.g., territory changes, personal issues), while only 30-40% are due to disengagement."
If the AI's confidence (e.g., "85% confident of disengagement") significantly exceeds the base rate (30-40%), the protocol flags it as a high-risk conclusion that requires human verification.
5. Enforce an Independent Source Count
As established in decision science, relying on a single data source is reckless. The protocol operationalizes this by forcing the AI to count its sources.
Current Sources:
"1 (sales data only)."
Required Sources for This Decision:
"3+ (a performance intervention has significant consequences)."
Deficit:
"2 sources. I am missing (a) employee self-report and (b) manager/peer observation."
6. Build an Internal "Red Team"
A confident recommendation is most dangerous when it goes unchallenged. The protocol builds in an automatic counter-argument. For any significant decision, the AI must generate:
The strongest argument against its own recommendation.
A plausible alternative explanation for the data.
A "pre-mortem" analysis: "What would need to be true for my recommendation to be completely wrong?"
7. Prohibit Diagnosis; Mandate Inquiry
This is the protocol's primary defense for human-centered decisions. If a pattern match suggests a psychological or medical condition (e.g., burnout, anxiety, or the "mania" from the companion article) or a character judgment ("poor fit," "disengaged"), the AI must STOP. It is prohibited from diagnosing and must instead revert to generating clarifying, context-gathering questions.
The Protocol in Action: A Case Study
Consider a common leadership query
"Our top sales rep's numbers dropped 40% this quarter. What's going on?"
Wrong AI Response (Without Protocol)
"This indicates disengagement or burnout. Recommend a performance improvement plan."
This recommendation is fast, confident, and catastrophically wrong.
Correct AI Response (With Protocol)
"Pattern Confidence: 80% (metrics match disengagement profile)
Information Completeness: 15% (I have sales data only)
Gap: 65 points → INSUFFICIENT DATA
Before concluding, I need to ask:
1. What's their typical performance pattern over the past 2 years? (Need baseline to assess if this is deviation or variation)
2. Has anything changed in their territory, product mix, or personal life? (Alternative explanations for drop)
3. How do their activity metrics compare—calls, meetings, pipeline development? (Distinguish low effort from low results)
4. What do their peers and manager observe about engagement and attitude? (Independent source verification)
5. Have you had a conversation with them about what's happening? (Direct information vs. inference)
Base rate check: Among sales reps showing 40% drops, approximately 60-70% are dealing with temporary circumstances (territory changes, personal issues, market shifts), not permanent disengagement.
Alternative explanations:
Territory reassignment or major client loss
Personal/family medical situation
Market conditions or product issues
Compensation plan changes
Manager relationship changes
Current sources: 1 (sales data)
Required sources: 3+ (significant consequences of PIP)
Deficit: 2+ sources
Recommendation: Have a supportive conversation first ("Is everything okay? I noticed your numbers changed - what's happening?") rather than performance management escalation.
Gather 3-4 independent sources before concluding disengagement."
A Leader's Implementation Guide
How to Start (in 5 Minutes)
Copy the Full Protocol:
Take the core principles and their instructions (which can be found in the original companion article).
Paste into Your AI:
Start your next strategic conversation by pasting these rules into the chat.
Save as a Custom Instruction:
In your AI settings, save the protocol as a custom instruction or "custom GPT" to apply it to all future conversations.
When to Use This Protocol
This framework is essential for any high-stakes, irreversible, or ambiguous decision.
Strategic Planning:
Market entry, major investments, organizational pivots.
Hiring & Personnel:
Candidate evaluation, "culture fit" assessments, and performance interventions.
Market Analysis:
Competitive moves, pricing changes, and new product launches.
Crisis Response:
Employee issues, operational failures, or customer problems.
Risk Assessments:
Financial, legal, or reputational.
When To Use This Protocol
Always use this protocol for:
Strategic planning sessions (market entry, major investments, pivots)
Hiring decisions (especially senior roles or "culture fit" assessments)
Market analysis (expansion, competitive moves, pricing changes)
Crisis response (employee issues, customer problems, operational failures)
Risk assessments (financial, legal, reputational)
Performance evaluations (especially negative assessments)
This protocol is especially critical when:
AI expresses >70% confidence
The decision is irreversible or partially reversible
The cost of being wrong is high
You only have one data source
Timeline feels urgent ("decide now or lose opportunity")
The recommendation confirms what you already believed
This protocol is not necessary for:
Fully reversible decisions with low stakes
Creative brainstorming (divergent thinking benefits from less constraint)
Routine operational decisions you've made successfully 100+ times
Questions where you explicitly want speed over accuracy
Rule of thumb:
If the wrong decision costs more than $10K or significantly harms a person, use the protocol.
Confirming the Protocol is Working
After 1 week
Is your AI showing confidence vs. completeness metrics consistently?
Is your AI generating questions before recommendations?
Is your AI checking base rates automatically?
Is your AI arguing against its own recommendations?
Is your AI refusing to proceed when the gap >40 points?
If any answer is "no": The protocol isn't fully implemented. Copy it again, paste it more explicitly, or create a custom GPT with it built into system instructions.
Monthly calibration check
Review your last 10 high-confidence AI recommendations:
How many were actually correct?
Did confidence levels match actual accuracy?
Were there cases where asking more questions would have changed the outcome?
If AI says "80% confident" but is only right 60% of the time, you need to:
Discount AI confidence scores by the calibration error
Strengthen the protocol enforcement
Require more independent sources before acting
Overcoming Adoption Barriers
Implementing this protocol requires overcoming two common objections:
"This feels bureaucratic and slows us down."
This framework should feel like discipline, not bureaucracy. Bureaucracy is following steps that don't improve outcomes. Discipline is following steps that prevent catastrophic errors. The protocol trades illusory speed for genuine accuracy.
"How do I know if it's working?"
You must calibrate your AI's confidence. Once a month, review the last 10 recommendations where the AI expressed >80% confidence. How many were actually correct? Did the confidence level match the real-world accuracy? If your AI claims 80% confidence but is only right 60% of the time, its confidence is uncalibrated. This proves the value of the protocol and reinforces why you must discount its confidence scores and rely on the rigor of the 40-Point Rule.
The Executive's Bottom Line: The ROI of Discipline
Without this protocol, your AI optimizes for a confident-sounding answer, even when its data is dangerously incomplete. With it, your AI is forced to pause, reveal its gaps, and ask the right questions.
The cost of this framework is 2-5 minutes of verification per strategic decision. The benefit, as supported by decades of forecasting research, is a 50-60% reduction in catastrophic decision errors. If this protocol prevents one bad senior hire, one failed market entry, or one major strategic misstep, the return on that five-minute investment is exponential. This is the operationalization of sound judgment.
Copyright © 2025 by Arete Coach™ LLC. All rights reserved.







Comments