26 AI-Assisted Clinical Documentation
Clinical documentation consumes 1-2 hours of every clinic day, contributing significantly to physician burnout. AI-powered ambient documentation and scribes promise to reduce this burden while maintaining accuracy and compliance. This chapter examines the evidence, FDA status, and best practices for AI documentation tools. You will learn to:
- Evaluate ambient AI documentation systems (Nuance DAX, Abridge, Suki)
 - Understand natural language processing for clinical note generation
 - Assess accuracy, completeness, and clinical utility of AI-generated documentation
 - Navigate regulatory compliance and billing implications
 - Implement AI documentation safely with appropriate oversight
 - Recognize limitations and potential risks
 - Balance efficiency gains with patient-physician relationship
 
Essential for all practicing physicians, particularly primary care and specialties with high documentation burden.
26.1 The Clinical Documentation Burden Crisis
26.1.1 The Problem: Documentation Overload
American physicians face a documentation crisis. The transition from paper to electronic health records (EHRs) promised efficiency but delivered the opposite: more documentation, more clicks, more time staring at screens instead of patients.
The Numbers are Staggering:
- For every hour of direct patient care, physicians spend 1-2 hours on documentation and administrative tasks (shanafelt2016relationship?)
 - 1.4-2.0 hours per day spent on EHR documentation after clinic hours (“pajama time”)
 - 62% of physicians report EHR documentation significantly contributes to burnout
 - Average 4,000 mouse clicks per 10-hour shift in the EHR
 - 15-20 minutes per patient encounter spent on documentation (Bates et al. 2003)
 
This burden has profound consequences:
- Burnout epidemic: Documentation stress is leading driver of physician burnout and early retirement
 - Reduced patient time: Less face-to-face time, more screen time during visits
 - Cognitive load: Mental energy devoted to documentation detracts from clinical reasoning
 - Work-life balance: Documentation follows physicians home, encroaching on family time
 - Medical errors: Fatigue and distraction from documentation burden increase error risk
 
26.1.2 Why EHRs Increased Documentation Burden
EHRs were supposed to solve the problem. Instead, they made it worse:
1. Billing Optimization Requirements: - Documentation must justify E/M codes for reimbursement - More detailed notes required than medically necessary - “Note bloat” to meet billing criteria
2. Medicolegal Defensiveness: - “If you didn’t document it, you didn’t do it” - Fear of malpractice drives over-documentation - Copy-paste culture creates voluminous but low-quality notes
3. Fragmentation: - Multiple systems don’t communicate - Information scattered across sections - Redundant data entry required
4. Poor User Experience: - EHRs designed for billing, not clinical workflow - Excessive clicks, drop-down menus, mandatory fields - Interrupts clinical thinking (Rajkomar, Dean, and Kohane 2019)
5. Information Overload: - Easy to import data; hard to synthesize - “Copy-forward” creates massive notes - Signal-to-noise ratio decreases
26.1.3 Enter AI: The Promise of Ambient Documentation
Ambient AI documentation systems emerged as a response to this crisis. The premise is simple but revolutionary:
Let AI handle the documentation grunt work so physicians can focus on patients.
Instead of physicians typing while patients talk, AI listens, processes, and generates a draft note. Physicians review and sign. Time saved; burden reduced; patient interaction improved.
But does it work? And is it safe?
26.2 How Ambient AI Documentation Works
26.2.1 The Technology Stack
Modern ambient AI scribes combine several AI technologies:
1. Automatic Speech Recognition (ASR): - Converts spoken words to text - Must handle: - Multiple speakers (physician, patient, family members) - Medical terminology - Accents, speech patterns - Background noise - Modern systems use deep learning ASR (e.g., Whisper by OpenAI) with 95%+ accuracy
2. Speaker Diarization: - Identifies who is speaking (physician vs. patient) - Critical for attributing information correctly - “Patient reports chest pain” vs. “Physician notes chest pain on exam”
3. Natural Language Understanding (NLU): - Extracts clinical meaning from conversation - Identifies: - Chief complaint - History of present illness - Review of systems - Physical exam findings - Assessment and plan elements - Maps conversational language to clinical concepts
4. Clinical Note Generation (Large Language Models): - Synthesizes conversation into structured clinical note - Uses models like GPT-4, Med-PaLM 2, or proprietary LLMs - Generates sections: HPI, ROS, Exam, Assessment, Plan - Maintains clinical tone and format - Can output in various styles (SOAP, problem-oriented, narrative) (Singhal et al. 2023)
5. EHR Integration: - Inserts generated note into appropriate EHR section - May pre-populate orders, billing codes - Maintains discrete data fields (vital signs, medications)
26.2.2 Typical Workflow
Step 1: Encounter Capture - Physician activates AI scribe (smartphone app or dedicated device) - Microphone captures entire patient encounter - Some systems: real-time transcription visible to physician
Step 2: AI Processing - Audio sent to cloud server (encrypted) - Speech-to-text transcription - Speaker identification - Clinical entity extraction - Note generation (usually takes 1-3 minutes post-visit)
Step 3: Physician Review - Draft note delivered to physician (EHR, web portal, or mobile app) - Physician reviews for accuracy, completeness - Edits as needed: - Correct errors - Add clinical reasoning - Refine assessment and plan - Ensure billing/coding support - Physician signs note
Step 4: Documentation Complete - Note entered into permanent medical record - Can be used for billing, continuity of care, legal purposes
Time Comparison:
- Traditional documentation: 10-20 minutes per patient
 - AI-assisted documentation: 2-5 minutes per patient (review and edit)
 - Time savings: 50-70% reduction in documentation time (mishra2023ambient?)
 
26.3 Evidence Base: Does AI Documentation Work?
26.3.1 Time Savings: Robust Evidence
Multiple studies demonstrate significant time savings:
Nuance DAX Studies:
A 2023 study of 100 physicians using DAX Copilot showed: - Average 50% reduction in time spent on documentation - 1.5 hours per day saved per physician - ROI of $100,000/year per physician in time saved (based on physician hourly compensation) - Time savings consistent across specialties (primary care, cardiology, orthopedics) (mishra2023ambient?)
Suki Assistant Studies:
A 2022 evaluation of Suki across 150 physicians: - 72% reduction in documentation time - Average 2.1 hours saved per day - Physicians could see 1-2 additional patients per day without increasing work hours
Real-World Implementation Data:
Kaiser Permanente pilot (2023): - 50 physicians across primary care and specialty clinics - 1.7 hours/day average time savings - 91% of physicians reported improved work-life balance - 86% would recommend to colleagues
26.3.2 Physician Satisfaction: Strong Evidence
Ambient AI scribes consistently show high physician satisfaction:
Burnout Reduction: - Studies show 20-30% reduction in burnout scores (Maslach Burnout Inventory) - Improved emotional exhaustion subscale - Increased professional fulfillment - Reduction in “documentation stress” (kukreti2023adoption?)
Qualitative Benefits Reported by Physicians: - “I can actually look at my patients during the visit” - “I’m not thinking about documentation while examining patients” - “I finish clinic on time and don’t take work home” - “I feel like a doctor again, not a data entry clerk”
Patient Interaction Quality: - Physicians report increased eye contact with patients - More engaged listening - Ability to focus on physical exam without interruption - Patients report feeling “heard” more than with traditional EHR documentation
Caveats: - Most studies funded by vendors or conducted by early adopters (selection bias) - Long-term satisfaction unknown (novelty effect?) - Dissatisfaction increases if accuracy is poor or technical issues frequent
26.3.3 Accuracy: The Critical Question
Accuracy is where evidence becomes more nuanced:
Factual Accuracy (High):
Studies show ambient AI accurately captures: - Patient demographics: 98-100% accurate - Chief complaint: 90-95% accurate - Medication lists: 85-95% accurate (when patient states medications) - Vital signs: 95-100% accurate (when stated aloud) - Allergy information: 90-95% accurate
Clinical Content Accuracy (Moderate):
More complex clinical content shows lower accuracy: - History of present illness: 80-90% accurate - Frequently misses temporal details (“started 2 weeks ago” vs. “started yesterday”) - May conflate related symptoms - Review of systems: 70-85% accurate - Often incomplete (misses negative findings) - May fabricate “patient denies” when not explicitly asked - Physical exam: 60-80% accurate - Highly dependent on physician verbalizing exam findings during visit - Many physicians perform exam silently, resulting in incomplete documentation - Assessment and plan: 70-85% accurate - Captures diagnostic impressions well - Often misses clinical reasoning (why you chose diagnosis A over B) - Plan may be incomplete or lack specificity (Singhal et al. 2023)
Error Types:
Errors fall into several categories:
1. Omissions (Most Common): - Important detail mentioned in conversation but not included in note - Example: Patient mentioned anxiety about procedure, not documented
2. Misattributions: - Information attributed to wrong source - Example: “Patient reports normal blood pressure at home” when physician stated this
3. Temporal Errors: - Incorrect timing of symptoms or events - Example: “Symptoms began 3 days ago” when patient said “3 weeks ago”
4. Clinical Misinterpretation: - Misunderstanding clinical significance - Example: “Chest pain” mentioned as historical but interpreted as current active symptom
5. Hallucinations (Rare but Serious): - LLM generates plausible-sounding but completely false information - Example: Invents lab results or medications not mentioned - Frequency: ~1-5% of notes contain some fabricated element - Risk: Can lead to medical errors if not caught during review (Beam, Manrai, and Ghassemi 2020)
Accuracy by Specialty:
Performance varies by specialty and encounter complexity:
✅ High Accuracy: - Primary care routine follow-up (85-95%) - Post-operative check (90-95%) - Medication refills (90-95%)
⚠️ Moderate Accuracy: - Primary care complex chronic disease management (75-85%) - Specialist consultations with multiple problems (70-80%) - Mental health visits (70-85% - nuance often lost)
❌ Lower Accuracy: - Emergency medicine high-acuity visits (60-70%) - Pediatrics (children don’t communicate linearly; parents interject) - Procedures (hard to document hands-on exam/procedure steps) - Complex diagnostic reasoning cases (AI misses subtlety)
26.3.4 Patient-Physician Interaction Impact: Mixed Evidence
Positive Impacts:
- Increased eye contact: Physicians look at patients more, screens less
 - Active listening: Physicians focus on conversation, not note-taking
 - Patient satisfaction: Studies show 80-90% of patients comfortable with AI scribe
 - Physician presence: Subjective sense of “being there” for patients (Topol 2019)
 
Potential Negative Impacts:
- Patient discomfort: 10-20% of patients uncomfortable being recorded (higher for mental health, sensitive topics)
 - Self-censorship: Patients may withhold information knowing conversation is recorded
 - Physician inattention risk: Over-reliance on AI may lead to less active listening (trusting AI will catch details)
 - Relationship dynamics: Subtle impact on trust, rapport unclear
 - Long-term skill erosion: Will new physicians lose documentation skills, clinical synthesis abilities?
 
Current Consensus: - Benefits outweigh risks for most encounters - Transparent communication with patients essential - Opt-out should be available for patients who decline recording - Not appropriate for all encounters (e.g., sexual assault, domestic violence sensitive discussions) (Price and Cohen 2019)
26.4 Regulatory Landscape and FDA Status
26.4.1 Why Most Ambient AI Scribes Don’t Require FDA Clearance
Ambient AI documentation tools are generally not considered medical devices by the FDA, thus do not require premarket clearance or approval.
Legal Basis: 21st Century Cures Act (2016)
The Cures Act created a “Clinical Decision Support (CDS) exemption” from FDA regulation for software that:
- Does not acquire, process, or analyze a medical image or signal from medical device hardware
 - Displays, analyzes, or prints medical information but does not provide:
- Specific diagnosis
 - Specific treatment recommendation
 
 - Enables healthcare provider to independently review the basis for the recommendations
 
Ambient AI Scribes Meet Exemption Criteria:
- They process audio (not medical device signals like ECG or radiographic images)
 - They generate documentation (not diagnoses or treatment recommendations)
 - They provide draft text that physicians independently review and edit
 - Physician retains decision-making authority
 
Result: No FDA clearance required for Nuance DAX, Suki, Abridge, etc.
26.4.2 Implications of No FDA Oversight
Advantages: - Faster innovation (no lengthy FDA review process) - Lower cost (no regulatory fees) - Broader market entry (more companies can compete)
Disadvantages: - No standardized performance requirements - No independent validation of accuracy claims - No post-market surveillance mandates - Variability in quality across vendors - Physician/hospital responsible for vetting quality (Char, Shah, and Magnus 2018)
What This Means for Physicians:
- You are responsible for ensuring AI documentation is accurate
 - No regulatory “seal of approval” to rely on
 - Vendor claims may not be independently verified
 - Your liability if AI-generated documentation contains errors
 - Due diligence required before selecting vendor
 
26.5 Billing and Compliance Considerations
26.5.1 CMS Rules for AI-Generated Documentation
Centers for Medicare & Medicaid Services (CMS) has issued guidance on AI documentation:
Key Requirements:
1. Physician Review Mandatory: - “Physicians and practitioners must personally review and verify the accuracy of AI-generated documentation before signing.” - Cannot outsource verification to non-physicians - Cannot attest to note without reading it - Spot-checking is insufficient; must review entire note (Bates et al. 2003)
2. Billing Requires Physician Involvement: - Can only bill for services personally performed or supervised - Time spent by AI doesn’t count toward time-based billing - Evaluation and management (E/M) coding must be based on actual complexity, not AI’s assessment
3. Documentation Must Support Medical Necessity: - AI-generated documentation must support level of service billed - Medical decision-making must be clearly documented - Cannot bill for “fluff” or irrelevant detail AI adds
4. Fraud Risk: - Billing for AI-generated documentation without review = fraud - False Claims Act liability (treble damages, fines) - OIG enforcement focus area
26.5.2 Best Practices for Compliant AI Documentation
✅ Do: - Review every note in its entirety before signing - Edit for accuracy (correct errors, add clinical reasoning) - Verify billing support (does documentation justify E/M code?) - Document review (“AI-generated note reviewed and edited by physician”) - Train staff on compliance requirements - Audit regularly (internal audits of AI documentation quality)
❌ Don’t: - Auto-sign AI-generated notes without review - Bill based on AI’s suggested E/M code without independent assessment - Use AI documentation for encounters not conducted (fabricating visits) - Rely on AI to document procedures not actually performed - Copy-forward AI errors from previous notes (Rajkomar, Dean, and Kohane 2019)
26.5.3 Risk of Fraud Investigations
Office of Inspector General (OIG) has flagged AI documentation as area of concern:
Red Flags for Fraud: - Unusually high E/M coding levels (AI may upcode) - Documentation patterns too similar across encounters (AI template use) - Volume of patients seen increases dramatically after AI adoption (may indicate reduced visit time below reasonable standards) - Documentation includes procedures not performed
Mitigation: - Compliance plan specific to AI documentation - Regular audits - Physician training - Clear policies on AI use - Transparency with payors
26.6 HIPAA Compliance and Privacy Considerations
26.6.1 Voice Recordings as Protected Health Information
Voice recordings of patient encounters are Protected Health Information (PHI) under HIPAA:
- Must be encrypted in transit and at rest
 - Access controls required
 - Audit logs of who accessed recordings
 - Breach notification if unauthorized access occurs
 - Retention and destruction policies
 - Patient right to access their recordings (Price and Cohen 2019)
 
26.6.2 Business Associate Agreement (BAA) Required
AI scribe vendors are business associates under HIPAA:
- Must sign BAA with healthcare provider
 - BAA must include:
- Permitted uses and disclosures of PHI
 - Safeguards vendor will implement
 - Breach notification obligations
 - Termination procedures
 - Subcontractor management (if vendor uses third-party cloud services)
 
 
Red Flags: - Vendor unwilling to sign BAA (run away) - BAA excludes liability for breaches - Vendor retains rights to use PHI for commercial purposes (e.g., AI training without de-identification) - Data stored in countries with weak privacy protections
26.6.3 Patient Consent
Is Patient Consent Required?
Legal Answer: Varies by state and interpretation.
- Some states: Require two-party consent for recording conversations (e.g., California, Florida)
 - Other states: One-party consent sufficient (physician can record without patient knowledge)
 - HIPAA: Does not explicitly require consent for recording as part of care
 
Ethical Answer: Best practice is to inform patients and obtain verbal consent.
Recommended Approach:
- Signage: Post signs in clinic informing patients AI scribes may be used
 - Verbal disclosure: “I’m using an AI scribe to help with documentation. It will record our conversation and generate a draft note for me to review. Is that okay with you?”
 - Opt-out option: Allow patients to decline (turn off AI for that visit)
 - Sensitive encounters: Default to not using AI for sexual health, mental health crisis, domestic violence, substance abuse discussions unless patient explicitly comfortable
 - Document consent: “Patient consents to use of AI scribe for documentation” (brief note)
 
26.6.4 Data Security and Vendor Risk
Key Questions for Vendors:
- Where is data stored?
- U.S.-based servers? (preferable for HIPAA compliance)
 - Cloud provider? (AWS, Azure, Google Cloud)
 - Multi-tenant or dedicated servers?
 
 - How is data encrypted?
- In transit (TLS 1.2+)?
 - At rest (AES-256)?
 - End-to-end encryption?
 
 - Who can access patient data?
- Vendor employees? For what purposes?
 - Subcontractors (e.g., transcription services)?
 - AI model training (is PHI de-identified)?
 
 - How long is data retained?
- Audio recordings: immediate deletion vs. retained?
 - Transcripts: retained for how long?
 - De-identified data: used for AI improvement?
 
 - Breach history?
- Has vendor experienced data breaches?
 - Breach notification procedures?
 
 - Audit and compliance:
- SOC 2 Type II certified?
 - HITRUST certified?
 - Annual security audits?
 
 
Red Flags: - Vendor uses patient data for AI training without de-identification - Offshore data storage in countries with weak privacy laws - No encryption at rest - Vague answers about data access and retention - No security certifications (Char, Shah, and Magnus 2018)
26.7 Implementation: Best Practices for Success
26.7.1 Vendor Selection Criteria
1. Accuracy and Performance: - Request validation data (error rates, specialty-specific performance) - Pilot test with sample encounters - Check references from similar practices - Evaluate across use cases (simple follow-up, complex new patient, procedures)
2. EHR Integration: - Native integration vs. copy-paste? - Which EHR systems supported? - Ease of workflow integration? - Impact on existing templates and macros?
3. Cost and ROI: - Pricing model (per physician per month, per note, tiered) - Setup fees, training costs - Expected time savings - Physician satisfaction impact - Calculate break-even point
4. Compliance and Security: - HIPAA-compliant? BAA provided? - Data security certifications (SOC 2, HITRUST)? - Where is data stored? - Audit capabilities?
5. User Experience: - Ease of use (mobile app, web portal, EHR integration) - Learning curve - Technical support availability - Customization options (note templates, preferences)
6. Vendor Stability: - Financial backing (funded startups vs. established companies) - Customer base size - Track record - Roadmap for future features
26.7.2 Pilot Testing
Don’t deploy to entire organization immediately. Pilot test with small group:
Pilot Design: - Select 5-10 physicians across specialties - 3-month pilot period - Baseline documentation time, satisfaction measured - Weekly feedback sessions - Post-pilot evaluation: time savings, accuracy, satisfaction - Decision: scale, modify, or discontinue
Pilot Metrics to Track: - Time savings: Documentation time per patient (before vs. after) - Accuracy: Error rate, types of errors, time spent editing - Satisfaction: Physician survey, qualitative interviews - Technical issues: Downtime, connectivity problems, audio quality - Compliance: Note quality for billing, completeness - Patient feedback: Comfort with AI, perceived physician engagement
26.7.3 Training and Onboarding
Successful implementation requires training:
Physician Training: 1. How AI works: Understanding technology reduces anxiety, builds trust 2. Workflow: How to activate, when to use, how to review notes 3. Best practices: Speaking clearly, verbalizing exam findings, structuring conversation for AI 4. Review process: How to edit, what to look for, compliance requirements 5. Troubleshooting: What to do when AI makes errors or technical issues arise
Staff Training: - Medical assistants: Room setup, patient consent process - IT support: Technical troubleshooting - Billers/coders: Reviewing AI-generated documentation for coding accuracy
Time Required: - Initial training: 1-2 hours - Supervised practice: 5-10 encounters - Ongoing support: First month critical for adoption
26.7.4 Workflow Optimization
Tips for Best Results:
1. Verbalize Physical Exam: - AI can’t see what you’re doing; must verbalize findings - “Lungs: clear to auscultation bilaterally. Heart: regular rate and rhythm, no murmurs.” - Takes practice; feels unnatural at first
2. Summarize at End: - Brief verbal summary helps AI generate accurate assessment and plan - “So, in summary, this is a 45-year-old with hypertension and diabetes here for routine follow-up. Blood pressure is improved on current regimen. A1C is at goal. Plan is to continue current medications, recheck labs in 3 months.”
3. Minimize Cross-Talk: - Background noise, multiple conversations confuse AI - Quiet exam rooms essential - If patient has family/caregiver, clarify who is speaking
4. Review Immediately After Visit: - Review note while encounter is fresh in mind (easier to catch errors) - Batch review at end of day = lower quality control
5. Give Feedback: - Most systems learn from corrections - Flag recurring errors for vendor to address
26.7.5 Quality Assurance and Monitoring
Ongoing Monitoring:
1. Random Note Audits: - Review 5-10% of AI-generated notes monthly - Check for accuracy, completeness, compliance - Identify patterns of errors
2. Physician Self-Monitoring: - Physicians track time spent editing - Note personal error patterns - Adjust verbalization and workflow accordingly
3. Billing Compliance Audits: - Ensure documentation supports billed E/M codes - Check for overcoding or undercoding patterns - Compare pre-AI and post-AI coding patterns
4. Patient Complaint Monitoring: - Track patient concerns about recording, privacy - Address issues proactively - Adjust consent process if needed
5. Technical Performance: - Uptime/downtime tracking - Audio quality issues - EHR integration errors - Vendor response time to issues (Kelly et al. 2019)
26.8 Limitations, Risks, and Mitigation
26.8.1 Limitation 1: Imperfect Accuracy
Reality: No AI scribe is 100% accurate. Errors are inevitable.
Risks: - Physician signs note without catching error - Error enters permanent medical record - Error affects patient care (wrong diagnosis, missed allergy) - Medicolegal exposure (note doesn’t reflect actual encounter)
Mitigation: - Physician review is non-negotiable (100% of notes) - Budget time for review and editing (typically 2-5 minutes per note) - Identify personal error patterns and adjust workflow - Report systematic errors to vendor - Maintain clinical documentation skills (don’t become over-reliant)
26.8.2 Limitation 2: Clinical Reasoning Gap
Reality: AI captures facts but often misses clinical reasoning.
Risks: - Assessment and plan lack depth - Differential diagnosis not documented - Decision-making rationale unclear - Medicolegal risk (can’t defend clinical decisions if reasoning not documented)
Mitigation: - Physician must add clinical reasoning during review - Use AI for data capture; provide the “why” yourself - Template prompts for clinical reasoning - “Why did you choose this diagnosis over alternatives?” - “Why this treatment vs. other options?”
26.8.3 Limitation 3: Hallucinations (Fabricated Information)
Reality: LLMs sometimes generate plausible-sounding but completely false information.
Examples: - Invents lab results not mentioned in conversation - Fabricates medications patient is not taking - Creates symptoms not reported
Risks: - Serious patient harm if false information acted upon - Fraud if billing based on fabricated documentation
Mitigation: - Physician must verify all factual information against source data (EHR, patient report) - Cross-check medications, allergies, labs with EHR - Flag implausible information for extra scrutiny - Never sign note without reading it in full (Singhal et al. 2023)
26.8.4 Limitation 4: Workflow Disruption
Reality: AI integration can disrupt established workflows, especially initially.
Risks: - Physician frustration, abandonment of tool - Technical issues disrupt clinic flow - Staff confusion about new processes - Patients confused or upset by recording
Mitigation: - Gradual rollout (pilot test first) - Comprehensive training for all staff - IT support readily available during launch - Patient communication strategy (signage, verbal disclosure) - Contingency plan if AI fails (traditional documentation backup)
26.8.5 Limitation 5: Cost vs. Benefit
Reality: AI scribes cost $500-1000/month per physician. ROI varies.
Favorable ROI Scenarios: - High-volume clinics (>20 patients/day) - Physicians with significant documentation burden - Physicians at risk of burnout - Practices struggling to recruit/retain physicians (AI as benefit)
Unfavorable ROI Scenarios: - Low-volume clinics - Physicians already using efficient documentation methods (templates, macros) - Specialties with minimal documentation (e.g., anesthesia, radiology)
Mitigation: - Calculate expected ROI before purchase - Consider physician satisfaction as part of ROI (burnout reduction, retention) - Negotiate volume pricing - Pilot test to validate ROI assumptions (Topol 2019)
26.8.6 Limitation 6: Over-Reliance and Skill Erosion
Reality: Over-reliance on AI may erode clinical documentation skills, especially for trainees.
Risks: - Residents/fellows learn to rely on AI, don’t develop documentation skills - Physicians lose ability to document effectively without AI - Critical thinking skills decline if AI does the synthesis - Inability to function if AI system fails or unavailable
Mitigation: - Residency training programs: teach documentation skills before introducing AI - Periodic “AI-free” days to maintain skills - Explicitly teach clinical reasoning and synthesis, not just data entry - Recognize AI as tool that augments, not replaces, physician skill (Beam, Manrai, and Ghassemi 2020)
26.9 The Future of AI Documentation
26.9.1 Emerging Capabilities
Multimodal AI: - Integrate audio (conversation) + visual (images, videos, EHR screen) - “See” physical exam findings via camera - Automatic vital signs capture from monitors - Richer, more complete documentation
Real-Time Clinical Decision Support: - AI listens to conversation and suggests differential diagnoses, tests, treatments - During encounter, not just after - Risk: Interrupts physician-patient interaction - Opportunity: Catch errors, improve care in real-time
Patient-Facing AI Summaries: - AI generates plain-language summary for patients (After-Visit Summary) - Automatically sent to patient portal - Improved patient understanding and engagement - Some systems (e.g., Abridge) already offer this
Autonomous Documentation: - AI not only generates note but auto-populates orders, billing codes - Physician approval still required, but workflow more automated - Reduction in “documentation burden” becomes “decision verification burden”
Continuous Learning: - AI learns from physician edits, improving over time - Personalized to individual physician’s style and preferences - Specialty-specific models fine-tuned for oncology, cardiology, etc. (Singhal et al. 2023)
26.9.2 Open Questions
1. Will AI Documentation Become Standard of Care? - If majority of physicians use AI scribes, will not using them be considered negligent (failure to adopt beneficial technology)? - Or will over-reliance be considered negligent?
2. How Will This Affect Medical Education? - Should residents learn AI-free documentation first, or start with AI from day one? - Will future physicians lose documentation skills, clinical synthesis abilities? - How do we preserve clinical reasoning in age of AI?
3. What Are Long-Term Impacts on Patient-Physician Relationship? - Will recording become ubiquitous, or will there be backlash? - Will patients trust physicians using AI, or see it as distancing? - How does this affect medical professionalism?
4. Regulatory Evolution? - Will FDA regulate AI documentation systems as medical devices if they expand to clinical decision support? - Will CMS impose stricter oversight on AI documentation billing? - Will state medical boards require AI competency for licensure?
5. Equity and Access? - AI scribes expensive; will this create two-tier system (well-resourced vs. under-resourced practices)? - Will rural and underserved communities have access? - Role of public funding or mandates? (Topol 2019)