The Audit-then-Automate Framework

Reliable AI for a contact center cannot be built without first understanding where the work actually breaks down. This document describes the methodology Flexbone uses to audit patient access operations, classify call workflows by complexity, and deploy voice AI agents trained on real operational data rather than assumptions.

Not everything should be automated

In most outpatient facilities, the patient access team handling phone calls is one of the largest operational groups in the organization. Depending on size, this can range from a handful of representatives at a single-site practice to hundreds of staff spanning centralized call centers across a health system. Regardless of scale, these teams handle thousands of inbound and outbound calls each week—scheduling, rescheduling, cancellations, eligibility verification, prior authorization, referral coordination, billing questions, and everything in between. They are the front door to both care delivery and the revenue cycle.

The instinct with AI is to automate as much as possible, as fast as possible. But the reality inside most facilities is that no one actually knows which calls can be automated, which ones should not be, and which ones fall somewhere in between. Not every interaction is a simple, scripted exchange. Some calls require clinical judgment, emotional sensitivity, or real-time problem solving that no AI agent should handle alone. Others follow a rigid, predictable pattern that a well-trained agent can resolve faster and more consistently than a human—but you cannot tell which is which without looking at the data first.

Despite the scale and importance of this function, most outpatient facilities have limited visibility into how these teams actually perform. Call volume, handle times, resolution rates, eligibility accuracy, authorization completion, and patient friction points often go unmeasured because the data infrastructure to capture them simply does not exist. Leadership sees the staffing expense line item and occasionally hears complaints, but the granular picture of what happens across thousands of daily patient interactions—where calls stall, where patients drop off, where staff deviate from the script, where revenue leaks—remains opaque. Without that picture, any attempt to introduce automation is guesswork.

Patient access is where many downstream revenue cycle problems originate. A missed eligibility check becomes a denied claim six weeks later. A prior authorization that is not followed up on becomes a canceled procedure. A patient who cannot reach the call center becomes a no-show, leaving allocated OR time unused. These failures are not random—they are systemic, and they start at the phone.

How the Audit Works

Flexbone’s engagement model begins with an operational audit. Before recommending automation or workflow redesign, the Voice Room platform is deployed to capture, transcribe, and analyze every call across the organization.

The resulting data makes the invisible visible—quantifying operational friction, revenue leakage, and staffing inefficiencies at their source. Most organizations have never had a complete picture of what their patient access team actually does all day. They know the team is busy. They know patients sometimes complain about hold times. They may even track basic metrics like call volume or abandonment rate. But none of that tells you what is happening inside the calls themselves—where time is wasted, where processes break down, and where a well-built AI agent could take over entirely.

Before labeling anything as automatable, the process starts with ground truth: how the work is actually being done today. Every call is captured and transcribed, producing a verbatim record of how representatives navigate scheduling, eligibility checks, authorizations, and edge cases. The analysis does not rely on job descriptions or assumed workflows. It examines the exact conversational steps staff take—where they follow a predictable pattern, where they deviate, and where judgment is required. This distinction matters because it is the difference between deploying an AI agent that works and deploying one that frustrates patients and creates more work for staff than it saves.

1

Deploy Voice Room

Capture and transcribe every inbound and outbound call across all locations. No sampling, no estimation—every conversation is recorded and analyzed.

2

Classify Workflows

NLP clusters intent, topic, and conversational patterns from raw transcripts. Each call is tagged by category, complexity, and resolution path.

3

Map Complexity

Score each identified use case as easy, moderate, or complex based on decision-tree analysis of the actual conversational steps staff take.

4

Quantify Impact

Calculate hours, cost, and revenue at risk per workflow to prioritize automation targets. This is where the business case becomes concrete.

5

Build & Deploy

Purpose-built AI agents per use case, trained on real calls from the audit data—not generic scripts or hypothetical scenarios.

The classification is not theoretical—it comes directly from analyzing thousands of actual conversations and mapping the decision trees staff navigate in real time. This is also what separates Flexbone from vendors who demo a generic scheduling bot and call it AI transformation. The audit produces a detailed operational map of the organization before a single agent is built.


Multi-Location Ophthalmology Network

A six-location eye care network in a major metropolitan area with no systematic understanding of their patient access team’s operational health.

6
Locations audited
3,700+
Calls per week
35%
Scheduling volume
20%
Voicemail volume

This organization had been growing steadily—adding locations, hiring providers, expanding services—but their patient access operations had not scaled with them. Leadership knew the phones were busy. They knew patients sometimes waited too long. But they had no way to quantify what was actually happening across thousands of weekly calls, no way to tell which workflows were consuming the most staff time, and no way to measure whether the team was following the processes they had been trained on.

After deploying Voice Room, the audit revealed that appointment scheduling accounted for more than one third of total call volume, making it the single largest driver of staffing demand and operating cost. Every new patient visit, follow-up, reschedule, cancellation, referral intake, and procedural booking flows through this function. Because it sits at the intersection of clinical capacity and patient demand, even small inefficiencies compound quickly across thousands of interactions each month.

The voicemail problem

Voicemail represented roughly a fifth of total call volume, but its operational impact was far larger than that number suggests. The core issue was not the volume itself—it was the phone tag it created and the complete lack of visibility into response times.

When a patient calls and reaches voicemail, they leave a message and wait. The staff retrieves the message, replays it, tries to decipher the request, looks up the patient in the system, and calls them back. If the patient does not answer, the staff leaves their own voicemail, and the cycle continues. A task that should take two minutes on a live call can stretch across hours or days of back-and-forth. Multiply that by hundreds of messages per week and the labor cost becomes staggering—not because any single message is hard to process, but because the workflow itself is broken.

Before the audit, the organization had no visibility into how long it took staff to return patient calls. They did not know if a voicemail sat for an hour or two days before someone acted on it. They did not know how many calls it took to actually reach the patient back. They did not know how many patients gave up waiting and either no-showed their appointment or found another provider entirely. The audit quantified all of this for the first time, and the numbers were worse than anyone expected.

Critical finding: The audit revealed voicemail boxes reaching capacity and rejecting messages. Patients heard “the mailbox is full and cannot accept any messages at this time” and disconnected. Those patients did not reschedule. Some became no-shows. Others found another provider entirely—revenue that walked out the door silently, with no one in the organization even aware it happened.

SOPs exist, but no one knows if they are being followed

The organization had invested real time and effort into building standard operating procedures for their patient access team. They had scripts for how to greet patients, protocols for how to handle insurance questions, and escalation paths for clinical inquiries. Supervisors had trained new hires on these processes. The assumption was that once someone was trained, they followed the playbook.

But the audit told a different story. When you can listen to every call—not just the ones a supervisor happens to overhear—you find out very quickly how often staff deviates from the SOP. Some representatives skipped the insurance verification step entirely. Others gave inconsistent information about co-pays and patient responsibility. A few handled scheduling requests differently depending on the time of day or how busy they were, creating an inconsistent patient experience across locations.

The problem is not that the SOPs were bad. The problem is that there was no way to monitor compliance at scale. A supervisor can sit with a representative and listen to a handful of calls per week. That is a rounding error against the thousands of calls the team handles. With Voice Room, every single call is captured, transcribed, and analyzed against the SOP framework. The organization can now see, in real data, which processes are being followed, which ones are being ignored, and where retraining or automation makes the most sense.

Audit findings by call category

CategoryVolumeKey Finding
Scheduling35%Single largest driver of staffing cost. Most calls follow a predictable, rules-based sequence with minimal variation.
Voicemail20%Disproportionate labor burden. Fragmented async workflow with no visibility into response times. Phone tag creating multi-day resolution cycles.
Insurance & Eligibility18%Verification gaps at intake create denied claims 4–6 weeks downstream. Most steps are deterministic and system-queryable.
Billing & Payments12%High repeat-call rate. Patients calling back about the same balance or statement multiple times due to inconsistent explanations.
Referrals & Auth10%Most time-intensive per call. Payer-specific workflows require different documentation per carrier.
Clinical & Other5%Requires human judgment. AI can triage and collect context before handoff to clinical staff.

In appointment scheduling alone, the workflow-level analysis identified over 500 fully automatable calls per day—thousands per week. The human labor and lost revenue this friction creates are measurable and directly impact a facility’s capacity for growth. But the number only becomes actionable once you know which specific calls those are, what conversational steps they involve, and how staff currently handles them. That is what the audit produces.

AI agents that work cannot be built without first understanding the bottlenecks of a facility. The audit is not a preliminary step—it is the foundation that determines what gets built, how it gets trained, and whether it actually reduces operational load or just adds noise.

Workflow Classification by Complexity

Every call transcript is analyzed against the actual conversational steps staff take. The classification determines what gets automated, what gets augmented, and what stays human.

Easy
85–95%
AI handling rate
Scripted, single-action flows. Transcripts show a consistent, rules-based sequence with minimal variation. The representative verifies identity, checks availability, offers standard slots, books the appointment, and closes. Fully automatable through voice AI, patient portal, or SMS.
Moderate
60–85%
AI handling rate
Multi-step but mostly deterministic. Calls follow a structured flow with occasional decision points. AI authenticates, captures intent, and completes standard scenarios, with automatic escalation when inputs fall outside predefined guardrails.
Complex
25–55%
AI handling rate
Real-time problem solving, policy interpretation, clinical nuance, or emotionally sensitive conversations. These require human judgment and remain human-led. AI triages, collects context, and hands off with full attribution of the issue.
In the ophthalmology network audit, 60% of total staff labor fell into the “Easy” bucket—simple, fully scriptable calls. Even with conservative automation estimates, these alone represented over $330,000 in annualized savings, freeing the team to focus on complex cases and patient retention.

Call Evidence

Actual call excerpts from Voice Room audits, demonstrating how workflows are classified and automation readiness is assessed from real conversations.

SchedulingEasy1 min 15 sec
“I need to reschedule my appointment from Thursday to sometime next week. Any morning slots available?”
Assessment: Standard identity-verify → check-availability → rebook flow. Resolved in under 90 seconds. No judgment required. Fully automatable.
BillingEasy1 min 44 sec
“I need a receipt for my HSA, and I can’t figure out how to get it through the portal.”
Assessment: Simple document lookup and email send. Patient had the receipt within two minutes. No reason this requires a human agent.
EligibilityModerate5 min 23 sec
“I just switched insurance plans and I’m not sure if my new carrier covers this procedure. Can someone check before I come in?”
Assessment: The representative explained the same eligibility verification process three separate times. An AI agent delivers a consistent explanation every time and queries the payer system directly—no frustration on either end.
ClinicalModerate0 min 45 sec
“I’ve been having issues with my vision since the procedure and I don’t know if I should come in or wait for my follow-up.”
Assessment: Correctly escalated to clinical staff within 45 seconds. AI can triage clinical questions and route to the right specialist instantly—no hold time, full context transferred.
VoicemailEasyAsync
“Hi, this is [name], I need to cancel my appointment on the 15th. Please call me back at…”
Assessment: Voicemail required manual retrieval, replay, and interpretation before any action could begin. AI transcribes, extracts intent and identifiers, and routes directly to the appropriate workflow queue—no human replay needed.

Impact Calculator

Adjust operating parameters to estimate the labor and cost impact of automating patient access workflows identified through an audit.

3,000
4.2 min
$22/hr
54%
54% is the median observed across audits for calls classified as fully automatable.
Hours Automated / Week
Monthly Savings
Annualized Savings
Current Annual Labor Cost
volume × AHT × rate × 52 wks
With Automation

Measurement Before Deployment. Monitoring After.

The audit data does not just inform what gets built—it becomes the training data and the ongoing performance benchmark.

Voice AI targets the specific call types that consume the most hours without requiring human judgment. Eligibility automation addresses the verification gaps that cause downstream denials. Prior authorization agents tackle the payer-specific bottlenecks that delay care and frustrate clinical staff. Each capability earns its place by solving a problem that is continuously monitored and improved upon.

The data from the actual organization serves not only as the foundation to understand what is possible and what to prioritize, but as the actual training data to deploy AI agents that work alongside existing staff rather than against them. When calls fall outside the automated scope, the AI agent transfers the caller directly to staff with full context—no dead ends, no dropped calls. The handoff is seamless because the agent already knows what the call is about, what the patient has said, and which staff member is best equipped to handle it.

This is where the SOP compliance monitoring becomes a force multiplier. The same infrastructure that powers the initial audit continues running after agents are deployed. Every call—whether handled by AI or by staff—is still captured, transcribed, and analyzed. The organization now has a continuous feedback loop: they can see which agents are resolving calls successfully, which ones are escalating too often, and which staff members are deviating from the protocols that were designed to prevent errors. Over time, this data compounds into an increasingly detailed operational map that makes every subsequent optimization faster and more targeted.

1 week
Typical audit timeline
1,000+
Calls reviewed per audit
40+
Use cases identified
28 days
Improvement cadence

As new call patterns emerge, new agents are trained and existing ones refined. Every 28-day cycle, Voice Room re-analyzes the data, measures resolution rates, and identifies where the system is underperforming. This is not a one-time deployment—it is a continuously improving operational layer that gets better the longer it runs. The first audit tells you what to build. The ongoing monitoring tells you what to build next, what to retrain, and where the organization is leaving money on the table.

For the ophthalmology network, the first cycle after deployment surfaced edge cases the initial audit had not anticipated—specific insurance plan variations, scheduling conflicts unique to certain providers, and patient questions about procedures that required more nuanced responses. Each of those became a targeted improvement. The system learned, adapted, and the resolution rate climbed. That is the compounding return of starting with real data instead of assumptions.

Start with the audit.

Within one week, a Flexbone Voice Room audit delivers a complete picture of patient access friction and identifies the AI transformation roadmap for an organization.

Book an Audit