Skip to content
  • About
  • Contact
  • Contribute
  • Book
  • Careers
  • Podcast
  • Recommended
  • Speaking
  • All
  • Physician
  • Practice
  • Policy
  • Finance
  • Conditions
  • .edu
  • Patient
  • Meds
  • Tech
  • Social
  • Video
    • All
    • Physician
    • Practice
    • Policy
    • Finance
    • Conditions
    • .edu
    • Patient
    • Meds
    • Tech
    • Social
    • Video
    • About
    • Contact
    • Contribute
    • Book
    • Careers
    • Podcast
    • Recommended
    • Speaking

Physicians must lead the vetting of AI

Saurabh Gupta, MD
Tech
November 11, 2025
Share
Tweet
Share

Walking out of the cath lab one night after taking care of a patient with an acute myocardial infarction, a thought crossed my mind. In cardiology, we would never deploy a new device without vetting it. Before a stent ever touches a coronary artery, it undergoes bench testing, animal trials, and human studies to prove safety and efficacy. This process typically takes years, and in some cases, decades. Data points and design decisions are validated and scrutinized. Yet many artificial intelligence systems influencing medical decisions today lack that rigor. From triaging chest pain in the ED to interpreting echocardiograms, from generating clinical notes to predicting readmission risk, AI now touches almost every corner of medicine. Tools such as ambient documentation and diagnostic support systems are becoming increasingly ubiquitous. However, while the technology has advanced at breakneck speed, our frameworks for validation remain archaic, outdated, or non-existent. The result is a widening gap between innovation and trust, and that gap is precisely where physicians must lead.

Why vetting AI matters more than ever

Earlier this year, I spoke at the American College of Cardiology’s Board of Governors meeting about the critical need for structured vetting of AI in clinical medicine. Because here’s the truth: Vetting doesn’t slow innovation; it makes it safe, ethical, and reproducible. Without it, enthusiasm risks outpacing evidence. We should evaluate AI with the same discipline we apply to any clinical tool. What does it mean to evaluate this with clinical rigor, and what frameworks might we consider:

  • Utility: Is it just technology in search of a solution, or does the technology actually improve outcomes or workflow?
  • Technical robustness: Is it accurate and precise? Does it demonstrate reliability across diverse populations and conditions, or does it fail at the margins?
  • Ethical integrity: Are we actively testing for bias before deploying it?
  • Regulatory transparency: Do we understand its logic well enough to explain it to a patient, or to a jury?

Every new model should address these questions before entering clinical care. AI may be capable of analyzing patterns we can’t see. However, it should still meet the same evidentiary standards as any medical device or drug.

The clinician’s evolving role

I see AI as amplifying the clinician’s role rather than diminishing it. Clinicians are ideally positioned to ensure the integrity and relevance of the generated insights. That requires us to transition from being passive end-users to active clinical stewards of technology. When physicians participate early in dataset design, bias testing, and post-market surveillance, we not only protect patients but also help build better AI. Clinical context is the missing ingredient that many tech companies underestimate. We must ask vendors and developers:

  • What data trained this model, and does it reflect my patient population?
  • How does it perform on populations like mine, not just on average?
  • What is its false-positive rate, and how do I verify its outputs?
  • When it fails, how will I know?

If we can’t answer these questions confidently, we shouldn’t use the tool. Physicians are the last line of defense between an algorithm’s confidence and a patient’s consequence.

From cath lab to courtroom: Applying medical rigor everywhere

The same principles of vetting clinical AI apply far beyond the hospital walls. In my work developing AI systems for high-stakes decision-making, our team of physicians, engineers, and legal experts faces these challenges daily. The challenges aren’t just technical; they are ethical. When an AI system organizes thousands of pages of medical records for a malpractice case or synthesizes evidence for peer review, accuracy isn’t optional. It’s foundational to fairness. We design our platforms with the same core principles we apply in medicine: traceability, validation, and human oversight. Every output links back to its source document, every finding can be audited, and every user maintains discretion over what is included in the record. We’ve learned that the same discipline of reasoning and transparent provenance we demand in clinical medicine should be applied in every domain where AI intersects with human judgment. Whether it’s a diagnostic decision in the ICU or a case review in a law firm, the principle remains the same: Trust comes from verification.

The real risk isn’t AI, it’s unvetted AI.

AI will make mistakes. So do we. The antidote isn’t fear; it’s accountability. That means continuous validation, bias detection, and human-in-the-loop oversight by design. It means demanding that we hold companies to the same exacting standards we have always held: sensitivity, specificity, positive predictive value, and negative predictive value, just as we expect from any diagnostic test. The most dangerous errors occur when neither the clinician nor the developer understands why the system failed. A biased dataset or a poorly generalized model can wreak havoc faster than any human could. The “black box” mindset must go. If an algorithm’s reasoning can’t be explained to a colleague, its utility in patient care should be questioned.

Leading with clinical integrity

The next generation of AI in professional domains will be judged by its credibility rather than by its complexity. That credibility begins with us. The question isn’t whether AI will transform medicine; it has already done so. The question is whether physicians will shape that transformation or be bystanders as the algorithmic race accelerates. As physicians, we already possess all the necessary skill sets to apply medical reasoning and transparency to the development, validation, and deployment of AI. We understand the stakes; after all, this is the reality of our everyday life. It behooves us to get it right. Responsible AI isn’t about slowing progress; it’s about ensuring that progress serves our patients well. When clinicians guide AI development and adoption, innovation aligns with ethics, and technology becomes an ally. AI doesn’t replace the physician. It tests whether we’re still willing to lead.

ADVERTISEMENT

Saurabh Gupta is an interventional cardiologist.

Prev

Physician entrepreneurship and financial freedom

November 11, 2025 Kevin 0
…
Next

South Carolina's CON repeal: an opportunity for doctors

November 11, 2025 Kevin 0
…

Tagged as: Health IT

Post navigation

< Previous Post
Physician entrepreneurship and financial freedom
Next Post >
South Carolina's CON repeal: an opportunity for doctors

ADVERTISEMENT

More by Saurabh Gupta, MD

  • Shopify’s meeting policy for health care: a game changer for providers and patients

    Saurabh Gupta, MD
  • Jimmy Carter brings attention to hospice care: the compassionate side of modern medicine

    Saurabh Gupta, MD

Related Posts

  • Migrant health in crisis: How we can lead the way in inclusive care

    Stephanie Dominic Berchmans, LMSW
  • Student loan forgiveness: a key step in achieving health equity for minority physicians and patients

    Katrina Gipson, MD, MPH
  • Social media: Striking a balance for physicians and parents

    Dawn Baker, MD
  • Why doctors must fight health misinformation on social media

    Olapeju Simoyan, MD
  • Family physicians unite at the U.S. Capitol, seeking congressional support for Medicare reform and health care transformation

    Tochi Iroku-Malize, MD, MPH, MBA, Sterling N. Ransone, Jr., MD, and Steven P. Furr, MD
  • Digital health equity is an emerging gap in health

    Joshua W. Elder, MD, MPH and Tamara Scott

More in Tech

  • Why voicemail in outpatient care is failing patients and staff

    Dan Ouellet
  • Building a clinical simulation app without an MD: a developer’s guide

    Helena Kaso, MPA
  • AI-enabled clinical data abstraction: a nurse’s perspective

    Pamela Ashenfelter, RN
  • Agentic AI in medicine: the danger of automating the doctor

    Shiv K. Goel, MD
  • Will AI replace primary care physicians?

    P. Dileep Kumar, MD, MBA
  • AI in medicine: Why it won’t replace doctors but will redefine them

    Tod Stillson, MD
  • Most Popular

  • Past Week

    • Health care as a human right vs. commodity: Resolving the paradox

      Timothy Lesaca, MD | Physician
    • My wife’s story: How DEA and CDC guidelines destroyed our golden years

      Monty Goddard & Richard A. Lawhern, PhD | Conditions
    • The gastroenterologist shortage: Why supply is falling behind demand

      Brian Hudes, MD | Physician
    • Why voicemail in outpatient care is failing patients and staff

      Dan Ouellet | Tech
    • Alex Pretti’s death: Why politics belongs in emergency medicine

      Marilyn McCullum, RN | Conditions
    • U.S. opioid policy history: How politics replaced science in pain care

      Richard A. Lawhern, PhD & Stephen E. Nadeau, MD | Meds
  • Past 6 Months

    • How environmental justice and health disparities connect to climate change

      Kaitlynn Esemaya, Alexis Thompson, Annique McLune, and Anamaria Ancheta | Policy
    • Will AI replace primary care physicians?

      P. Dileep Kumar, MD, MBA | Tech
    • A physician father on the Dobbs decision and reproductive rights

      Travis Walker, MD, MPH | Physician
    • What is the minority tax in medicine?

      Tharini Nagarkar and Maranda C. Ward, EdD, MPH | Education
    • Why the U.S. health care system is failing patients and physicians

      John C. Hagan III, MD | Policy
    • Alex Pretti: a physician’s open letter defending his legacy

      Mousson Berrouet, DO | Physician
  • Recent Posts

    • A physician’s quiet reflection on January 1, 2026

      Dr. Damane Zehra | Conditions
    • AI censorship threatens the lifeline of caregiver support [PODCAST]

      The Podcast by KevinMD | Podcast
    • Demedicalize dying: Why end-of-life care needs a spiritual reset

      Kevin Haselhorst, MD | Physician
    • Physician due process: Surviving the court of public opinion

      Muhamad Aly Rifai, MD | Physician
    • Spaced repetition in medicine: Why current apps fail clinicians

      Dr. Sunakshi Bhatia | Physician
    • When the doctor becomes the patient: a breast cancer diagnosis

      Sue Hwang, MD | Conditions

Subscribe to KevinMD and never miss a story!

Get free updates delivered free to your inbox.


Find jobs at
Careers by KevinMD.com

Search thousands of physician, PA, NP, and CRNA jobs now.

Learn more

Leave a Comment

Founded in 2004 by Kevin Pho, MD, KevinMD.com is the web’s leading platform where physicians, advanced practitioners, nurses, medical students, and patients share their insight and tell their stories.

Social

  • Like on Facebook
  • Follow on Twitter
  • Connect on Linkedin
  • Subscribe on Youtube
  • Instagram

ADVERTISEMENT

ADVERTISEMENT

  • Most Popular

  • Past Week

    • Health care as a human right vs. commodity: Resolving the paradox

      Timothy Lesaca, MD | Physician
    • My wife’s story: How DEA and CDC guidelines destroyed our golden years

      Monty Goddard & Richard A. Lawhern, PhD | Conditions
    • The gastroenterologist shortage: Why supply is falling behind demand

      Brian Hudes, MD | Physician
    • Why voicemail in outpatient care is failing patients and staff

      Dan Ouellet | Tech
    • Alex Pretti’s death: Why politics belongs in emergency medicine

      Marilyn McCullum, RN | Conditions
    • U.S. opioid policy history: How politics replaced science in pain care

      Richard A. Lawhern, PhD & Stephen E. Nadeau, MD | Meds
  • Past 6 Months

    • How environmental justice and health disparities connect to climate change

      Kaitlynn Esemaya, Alexis Thompson, Annique McLune, and Anamaria Ancheta | Policy
    • Will AI replace primary care physicians?

      P. Dileep Kumar, MD, MBA | Tech
    • A physician father on the Dobbs decision and reproductive rights

      Travis Walker, MD, MPH | Physician
    • What is the minority tax in medicine?

      Tharini Nagarkar and Maranda C. Ward, EdD, MPH | Education
    • Why the U.S. health care system is failing patients and physicians

      John C. Hagan III, MD | Policy
    • Alex Pretti: a physician’s open letter defending his legacy

      Mousson Berrouet, DO | Physician
  • Recent Posts

    • A physician’s quiet reflection on January 1, 2026

      Dr. Damane Zehra | Conditions
    • AI censorship threatens the lifeline of caregiver support [PODCAST]

      The Podcast by KevinMD | Podcast
    • Demedicalize dying: Why end-of-life care needs a spiritual reset

      Kevin Haselhorst, MD | Physician
    • Physician due process: Surviving the court of public opinion

      Muhamad Aly Rifai, MD | Physician
    • Spaced repetition in medicine: Why current apps fail clinicians

      Dr. Sunakshi Bhatia | Physician
    • When the doctor becomes the patient: a breast cancer diagnosis

      Sue Hwang, MD | Conditions

MedPage Today Professional

An Everyday Health Property Medpage Today
  • Terms of Use | Disclaimer
  • Privacy Policy
  • DMCA Policy
All Content © KevinMD, LLC
Site by Outthink Group

Leave a Comment

Comments are moderated before they are published. Please read the comment policy.

Loading Comments...