AI in Healthcare:
Value vs. Hype
A Practical Framework for Evaluating AI in Your Practice
Session Materials
The take-home framework is designed to be printed and kept at your desk. The slides include all cases, data, and references from the session.
What This Session Covers
AI tools are increasingly introduced into clinical practice — often without clear evidence of effectiveness, limitations, or failure modes. Physicians are frequently asked to evaluate these systems without a structured framework for assessing clinical value or implementation readiness. This session reviews current evidence on AI in clinical medicine, including documented successes, known failures, and the role of AI in clinical decision-making. Participants receive a practical five-question framework to guide evaluation and responsible adoption of AI tools within clinical practice.
AI in Healthcare by the Numbers
The market has moved faster than governance. Every physician is now encountering AI in their workflow — often without knowing it.
Healthcare Is Not a Monolith
AI readiness depends entirely on which part of healthcare you are talking about. Conflating the two is where most AI failures begin.
Not Everything Called "AI" Is the Same Thing
Six layers, each doing something different, each carrying different risk. When a vendor says "we use AI" — ask which layer. The answer changes everything about how you should evaluate the claim.
Risk: Moderate — Performance depends heavily on training data quality and whether local validation was performed
Risk: Moderate-High — Performance gaps in edge cases can be undetectable without prospective validation in your environment
Risk: Lower for flagging/triage functions — increases significantly for autonomous diagnostic interpretation
Risk: High when applied to clinical reasoning — proven effective only in bounded operational tasks like documentation
Risk: Moderate — Better than standalone LLMs, but only as reliable as the knowledge base it references
Risk: Experimental — Governance, liability, and patient safety frameworks do not yet exist for most agentic healthcare use cases
After This Session, You Will Be Able To…
Mapped to ACGME competency domains · 1.0 AMA PRA Category 1 Credit™
Evaluate AI tools through the lens of workflow integration, data integrity, human oversight, and end-user experience — for both clinicians and patients.
Patient Care & Practice-Based LearningApply a 5-question practical decision framework to assess AI vendor claims, identify implementation readiness gaps, and distinguish evidence-based tools from marketing-driven hype.
Systems-Based PracticeRecognize the role of physician leadership in AI governance — including how to advocate for clinician involvement in AI design, procurement, and institutional oversight.
Professionalism & CommunicationFormulate at least two concrete next steps to engage with AI governance, evaluation, or implementation within your own practice or institution.
Practice-Based Learning & ImprovementThe 5-Question Framework
Ask these before any AI adoption decision. Each maps to a domain in the AMA AI Tool Evaluation Guide — developed by 21 specialty societies, February 2026.
What specific, bounded problem does this solve?
If the answer is vague, stop there. Operational tasks work; broad clinical claims don't.
AMA Domain 01 — Clinical Use Case & UserWhere does the physician stay in the loop?
The final clinical decision must remain with a clinician. Always.
AMA Domain 03 — Risks & MitigationWhat was it trained on — and does it reflect OUR patients?
Local validation is not optional. The Epic Sepsis Model and the Optum algorithm both failed on this question — after national deployment.
AMA Domain 02 — Training & Validation DataHow were clinicians involved in building this?
Not just consulted at the end. Actually involved in design, testing, and iteration throughout the process.
AMA Domain 05 — Workflow Integration & MonitoringWhat does failure look like — and who's accountable?
If no one can answer this clearly, that IS your answer.
AMA Domain 04 — Effectiveness & Performance"AI tools that work solve narrow, bounded problems with clean data and a physician still in the loop. The ones that fail are solving a vendor's pitch deck."
What Actually Works
Every AI tool that works shares three traits: it solves one bounded problem, keeps a human in the loop, and runs on representative, validated data. These three cases meet all three criteria.
The Liability Gap
AI creates liability exposure in two directions simultaneously — and the legal standard of care is still being written in real time.
Start Here
Everything on this list can be done without a budget, IT approval, or anyone's permission.
Every Claim Has a Primary Source
All DOIs and external links are live. Grouped by source type.
DOI: 10.1001/jamanetworkopen.2025.34976 →
DOI: 10.1001/jamanetworkopen.2024.40969 →
DOI: 10.1001/jamainternmed.2021.2626 →
DOI: 10.1126/science.aax2342 →
DOI: 10.1093/jamiaopen/ooae133 →
clevelandclinic.org →
ecri.org →
ama-assn.org →
ama-assn.org →
ama-assn.org/steps-forward →
fda.gov →
fda.gov →
azmed.org →
manatt.com →
For Physicians Evaluating AI
Start with the AMA Guide. Everything else follows from it.
Accreditation: Presented in accordance with ACCME Standards for Integrity and Independence in Accredited Continuing Education.
Financial Relationships: Speaker has no relevant financial relationships with ineligible companies. No vendor payments, consulting arrangements, or speaker bureau relationships. Employer: Intermountain Health (non-commercial health system).
Commercial Support: No commercial support was accepted or solicited for this activity.
Content Standards: Content is nonpromotional, evidence-based, and free of commercial bias. All clinical claims cite peer-reviewed literature or primary source data.