What Happens When You Talk to AI
- The Most Important Thing to Understand
- Why This Matters Clinically
- How the Model Learned What It Knows
- The Training Data Problem
- How AI Reads Your Question
- Context Windows: What the AI Can See
- How the Response Is Built
- What This Means for Your Practice
- Before You Turn the Page
Why AI Sounds Confident But Gets It Wrong
- The Terminology Problem
- The Anatomy of a Confident Error
- Why “Just Verify It” Is Harder Than It Sounds
- The Five Failure Modes That Matter Most
- What the Research Shows
- The Ship at 2:47 AM, Continued
- Toward a Different Kind of Verification
- Before You Turn the Page
Where the Training Data Came From (And Why That Matters)
- The Corpus: What the Model Read
- What RLHF Actually Does (And What It Cannot)
- Bias: What the Data Does Not Contain
- The Forty-Drug Formulary Problem
- The Knowledge Cutoff and the Guideline Lag
- Why the Model Treats UpToDate and Reddit the Same
- Before You Turn the Page
The Six-Platform Study — What I Found When I Tested Clinical AI
- Why I Built This Study
- The Six Platforms
- The Rubric: 20 Items, 100 Points, One Standard
- The Results
- What the Clinical AI Tools Got Right
- Where the General LLMs Failed
- What the 2.4x Gap Actually Means
- The Benchmark Paradox
- What No Rubric Captures
- Why the Gap Exists: Architecture, Not Intelligence
- What This Means for Your Practice
- Study Limitations
- Before You Turn the Page
When AI Assumes Resources You Don’t Have
- The Default Environment
- The Maritime Stress Test
- Five Scenarios the Algorithm Gets Wrong
- The Evacuation Decision: What Algorithms Cannot Weigh
- Beyond the Ship: The Global Resource Gap
- Building AI That Knows Your Constraints
- What This Means for You
- Before You Turn the Page
The Failure Modes That Matter Most
- Drug Interaction Blindness
- Context Window Degradation
- Anchoring Bias Amplification
- Copy-Paste Liability: When AI Output Enters the Record
- Shadow AI: The Tools Nobody Talks About
- The Compound Effect
- Before You Turn the Page
A Clinician’s Framework for Evaluating AI Tools
- The Five Questions
- Red Flags in AI Product Marketing
- When the Institution Chooses for You
- Before You Turn the Page
Prompting Like a Clinician
- The Specificity Principle
- Structured Prompting Templates
- The Five Most Common Prompting Mistakes
- Eliciting Uncertainty
- Ship-Specific Prompting: Lessons from the Edge
- Documenting AI-Assisted Decisions
- Before You Turn the Page
The Regulatory Landscape — What Clinicians Need to Know
- What the FDA Has Done So Far
- The CDS Exemption — The Most Important Regulatory Concept You May Not Know
- The Gap Between Clearance and Safety
- Who Is Liable When AI Is Wrong?
- What Is Happening in Europe
- The American Patchwork
- What Clinicians Should Actually Do
- Before You Turn the Page
The Future of Clinical AI — From the Edge of Medicine
- The 80/20 Problem
- Architecture Is the Safety Layer
- The Reproducibility Problem
- What I Learned From Breaking Things
- The Physician’s Role in Building Better AI
- From the Edge of Medicine
- A Personal Note
Appendix A: Glossary of AI Terms for Clinicians
- How the Models Work
- Training and Alignment
- Errors and Failure Modes
- Architecture and Safety
- Regulation and Governance
- Evaluation and Benchmarks
Appendix B: AI Evaluation Checklist for Clinicians
- Regulatory Status Quick Check
- Documentation Template
Appendix C: Prompting Templates for Common Clinical Scenarios
- Template 1: Acute Pharmacological Management
- Template 2: Differential Diagnosis with Resource Constraints
- Template 3: Drug Interaction and Safety Review
- Template 4: Evacuation and Transfer Decision Support
- Template 5: Evidence Synthesis and Guideline Review
- Template 6: Adversarial Follow-Up (Universal)
- A Final Note on These Templates