Preface

AI as a Clinical Tool, Not a Revolution

Medicine has always been about pattern recognition: correlating symptoms with diagnoses, identifying subtle findings that distinguish one condition from another, processing vast amounts of information to make critical decisions under uncertainty.

Artificial intelligence is entering clinical medicine not as a revolution that will replace physicians, but as a new category of tool that requires the same critical evaluation we apply to any clinical intervention.

The challenge isn’t whether AI will transform medical practice. It already has. Algorithms interpret chest X-rays and detect pneumonia. Deep learning models analyze pathology slides and identify malignancies. Natural language processing extracts insights from electronic health records. Clinical decision support systems recommend evidence-based treatments.

But here’s what you rarely hear in the vendor pitches and conference keynotes: Most of these tools aren’t as good as advertised. Many fail in real clinical environments. Some have been withdrawn after causing harm. Even the successful ones require careful implementation, continuous monitoring, and honest acknowledgment of limitations.

The real challenge facing physicians is this: How do we critically evaluate these tools, integrate them responsibly into clinical workflow, understand their limitations and failure modes, navigate medico-legal implications, and ultimately decide which AI applications genuinely improve patient care? All while continuing to see patients, manage complex cases, and stay current with traditional medical literature.

This handbook addresses that challenge. Not by promising easy answers, but by providing honest, evidence-based guidance.

What This Handbook Is Not

This is not a computer science textbook. If you want to build neural networks from scratch or understand backpropagation algorithms, excellent technical resources exist elsewhere, and you don’t need that knowledge to use AI tools effectively in clinical practice.

This is not a manifesto claiming AI will solve medicine’s hardest problems. It won’t. Healthcare’s most challenging issues (health inequity, access barriers, burnout, administrative burden, the social determinants of health) are fundamentally systemic problems that no algorithm can fix alone.

This is not a catalog of futuristic possibilities 20 years away. This handbook focuses on what exists now, what has peer-reviewed clinical evidence, what physicians are actually using, what works (sometimes), what fails (often), and what you can realistically evaluate for your practice today.

This is not vendor-neutral in a false balance way. Where evidence clearly supports specific tools, I name them. Where tools have failed spectacularly despite marketing hype, I document those failures. Physicians deserve honest assessments based on clinical evidence, not diplomatic neutrality between good and bad applications.

What This Handbook Attempts

This is the resource physicians need when encountering AI tools in clinical practice, where traditional medical education hasn’t prepared them to evaluate these tools critically.

This handbook attempts to:

  • Translate without oversimplifying. AI is neither magic nor “just statistics”; understanding the middle ground matters
  • Ground everything in clinical evidence. Citations from JAMA, NEJM, The Lancet, Nature Medicine, BMJ, and specialty journals throughout
  • Acknowledge failures as prominently as successes. IBM Watson for Oncology’s failure teaches more than another diabetic retinopathy screening success story
  • Address specialty-specific applications. Pediatrics faces different AI challenges than radiology; both deserve dedicated attention
  • Focus on implementation realities. Technical performance metrics matter less than whether a tool integrates into actual clinical workflow
  • Center patient safety and medical liability. Algorithms don’t face malpractice suits; physicians do

A Note on Incompleteness and Uncertainty

This handbook will always be incomplete. Medical AI capabilities shift monthly. New applications emerge constantly. Regulatory frameworks evolve. Techniques that seemed promising become obsolete. Clinical trials report results that overturn yesterday’s conclusions.

This is a living document, not a static artifact. Better to have an honest, evolving resource than something polished but outdated before publication.

Where consensus exists, I state it clearly. Where evidence conflicts or gaps remain, I acknowledge uncertainty rather than manufacture false confidence.

This handbook prioritizes peer-reviewed evidence from major medical journals, FDA-cleared applications, and real-world clinical implementations over press releases, vendor whitepapers, and theoretical capabilities.

If you find evidence that contradicts what I’ve written (newer studies, clinical trials, systematic reviews, implementation failures I haven’t documented), please contribute. Medicine advances through evidence and honest correction, and this handbook should reflect our best current understanding, not my personal limitations.

On Reading This Book

You don’t need to read sequentially. Medical specialties differ substantially in AI maturity and applications. Jump directly to your specialty chapter. Use the search function. Read TL;DRs for quick orientation. Dive deep when evaluating tools for your practice.

Treat this as a field guide, not a textbook. Dog-ear the pages (digitally). Write in the margins (via GitHub discussions). Return to relevant sections when facing specific decisions. Share chapters with colleagues. Adapt frameworks to your institutional context. Question everything I’ve written, including the frameworks I’ve proposed.

Most importantly: Stay skeptical. Question vendor claims. Question the hype. Question the tools. Even question the limitations I’ve described. Medicine has suffered before from adopting technologies uncritically. We cannot afford to make the same mistake with AI.

Three reading approaches:

1. The Quick Scan (TL;DRs only): - Read chapter summaries for rapid orientation - Perfect for busy clinicians needing key takeaways - Use when evaluating whether a chapter addresses your immediate question - Best for: Attendings, department chairs, anyone triaging information

2. The Deep Dive (Full chapters): - Read complete chapters for detailed understanding - Includes clinical evidence, case studies, implementation details - Use when implementing AI tools or developing institutional policies - Best for: Informaticists, researchers, residency educators, early adopters

3. The Specialty Focus (Your chapters only): - Read foundations (Part I) + your specialty + implementation (Part III) - Most efficient path for specialty-specific guidance - Skip chapters outside your clinical domain - Best for: Practicing physicians in specific specialties

Remember: Demand prospective clinical trials, not just retrospective validation studies. Prioritize patient safety over efficiency gains. You remain legally and ethically responsible for clinical decisions, regardless of what any algorithm recommends.

How to Use the Chapter Summaries (TL;DRs)

Every chapter begins with an expandable Chapter Summary (TL;DR) section designed for rapid comprehension without sacrificing clinical accuracy.

What’s Inside Each TL;DR:

  • The Clinical Context: Why this matters for patient care
  • Key Evidence: Major studies and systematic reviews
  • What Actually Works: Evidence-based applications currently in clinical use
  • What Doesn’t Work: Documented failures and limitations
  • Clinical Bottom Line: The essential takeaway for practice
  • Medico-Legal Considerations: Liability and regulatory issues where relevant

Visual Indicators: - Safety concerns and important warnings - Evidence-supported applications - Failed implementations and what went wrong - Key clinical insights - Important data from trials and studies

The TL;DRs are not superficial summaries. They represent distillation of peer-reviewed literature into clinically actionable guidance. Many physicians have told me they use TL;DRs for 80% of their needs and read full chapters only when implementing specific tools.

That’s intentional. Your clinical time is precious. Use it wisely.

A Personal Note

I’m a physician-epidemiologist who spent the pandemic evaluating computational tools under high-stakes conditions, from genomic surveillance systems to outbreak prediction models. That experience taught me a framework for critical evaluation that applies directly to clinical AI.

I wrote this handbook because physicians deserve evidence-based, honest, practical guidance about AI tools that respects both clinical expertise and the complexity of real patient care. The decisions we make now about which tools to adopt and how to evaluate them will shape medical practice for decades.


Bryan Tegomoh, MD, MPH Berkeley, California January 2025

This site uses privacy-friendly analytics (Plausible) to track aggregate page views and improve the handbook. No cookies or personal data are collected.