top of page

Navigating Healthcare Compliance in the Age of AI: Lessons From a Major Case Study

  • Writer: Jessica Zeff
    Jessica Zeff
  • Dec 9, 2025
  • 3 min read

If you've been in healthcare compliance long enough, you start to recognize a pattern: the warning signs are always there. During our recent Compliance Deconstructed episode, Lorie Davis, Elvan Baker, and I took a hard look at a large healthcare organization—think tens of thousands of employees, sprawling facilities, advanced tech investments—that still managed to miss the mark on some of the most basic compliance expectations.



And it wasn’t just one misstep.


This organization’s failures spanned sterilization protocols, patient documentation, billing practices, and internal oversight. The ripple effect was massive—not just in financial penalties or regulatory attention, but in lost trust from patients and staff. In a world where AI is starting to influence clinical and administrative decisions, these foundational breakdowns raise a bigger question: How do we ensure technology doesn't outpace integrity?


Compliance and Quality: Two Sides of the Same Coin

Too often, compliance gets relegated to a checklist or policy binder. But as we talked about on the podcast, when sterilization isn't done properly or documentation is sloppy, patients are the first to feel it. The idea that compliance is “just the rules” misses the real point—compliance is how we deliver safe, consistent, and equitable care.


Some of the issues we discussed included:

 

  • Incomplete or missing documentation

  • Lapses in protocol adherence

  • Poor internal communication

  • Weak data governance

  • Unclear roles and responsibilities

 

Now, layer AI on top of that. Machine learning systems only work when the underlying data is clean, validated, and used in ethical ways. Garbage in, garbage out—except now, that garbage could affect an entire patient population.

 

Culture Eats Policy for Breakfast

This part hit home for all of us. The organization had all the trappings of compliance—policies, committees, budgets—but the culture didn't support it. People felt pressured to hit productivity metrics at the expense of process. Leaders turned a blind eye to known issues. And worst of all, reporting concerns was met with silence or subtle retaliation.


Sound familiar?


We’ve all been in rooms where performance metrics get more airtime than compliance indicators. But when staff don’t trust the system, they stop speaking up. And when that happens, risk becomes invisible—until it's on the front page of the Wall Street Journal.

 

The Billing Burden—and the Opportunity for AI

Billing is where the rubber meets the road for many compliance programs. It’s technical, fast-moving, and unforgiving. On the episode, we walked through several pain points:


  • Upcoding for higher reimbursement

  • Discrepancies between clinical documentation and billing entries

  • Lack of staff training on service codes

  • Missed documentation standards


AI promises to help here—automated audits, predictive error detection, coding assistants. But even AI can’t fix cultural or process gaps. You still need human oversight, role-based training, and rigorous validation processes to make it work safely.

 

Complex Organizations Need Simple Guardrails

In organizations with thousands of employees and dozens of sites, complexity is inevitable. But your compliance approach doesn’t need to be complicated—it needs to be clear. We explored strategies that help compliance leaders create visibility and control, even across sprawling systems:


  • Role-specific training tied to actual responsibilities

  • Real-time reporting systems with leadership follow-through

  • Escalation paths that actually go somewhere

  • Internal audits focused on high-risk areas, not just the easy-to-measure stuff

  • Policies that reflect how work really gets done—not how we wish it did


When these basics are solid, AI can add real value. Without them, it’s just more noise in an already overwhelmed system.

 

Looking Ahead: Ethical AI Requires Ethical Infrastructure

One takeaway from our conversation that I keep coming back to: We can’t separate tech from culture. You don’t get ethical AI from a vendor contract. You get it by building a workforce that knows how to ask, “Should we?”—not just “Can we?”


The healthcare organizations that will thrive in the AI era aren’t the ones chasing the newest tools. They’re the ones investing in cultures where compliance is part of daily decision-making. Where feedback isn’t feared. Where the people designing, validating, and using technology actually understand what’s at stake.

 

Final Thoughts

Compliance professionals aren’t just gatekeepers—we’re architects. We design systems that protect patients, support staff, and keep organizations aligned with their mission. As AI changes how care is delivered, our work becomes even more essential.

 


Watch the full episode above or listen everywhere you find your podcasts!



Comments


bottom of page