2026-03-14 · 3 min read
AI operating models in regulated environments: Beyond the pilot
Most organisations have run an AI pilot. Far fewer have shipped one to production in a regulated environment. The gap isn't technical — the models work fine. It's organisational, and honestly, it's almost always the same problems.
The experiment trap
Pilots are designed to succeed. They sit outside normal governance, with sympathetic stakeholders, cherry-picked use cases, and the shared understanding that nothing is permanent. When they work, there's a big presentation and a lot of excitement. When they fail, nothing really happened.
Production doesn't work like that. Production means real data, real users, real regulatory scrutiny, and actual accountability for what the system does when it gets something wrong. I've seen teams nail a pilot and then spend two years trying to get something into production. The prototype that impressed the steering committee and the system that runs in a live environment are basically different jobs.
What actually needs to exist
Here's the thing: the organisations failing at AI delivery aren't usually failing at AI. They're failing at the infrastructure around it.
Governance is the first blocker. Traditional risk and compliance frameworks weren't built for iterative development — they were built for waterfall. You need oversight that's rigorous but doesn't take six months to approve a model update. Getting those two things to coexist is genuinely hard.
Then there's accountability. Who owns it when the model returns a bad answer? In most organisations, this question has no clear answer, and everyone's comfortable leaving it that way. Until someone owns it, nothing significant ships. It's that simple.
The third gap is operational. Models in production need monitoring, feedback loops, and incident procedures — and most delivery teams treat this as someone else's problem. It isn't. Running AI in production requires a kind of operational thinking that most organisations genuinely haven't had to develop before. That's exciting, actually. There's a lot of space to figure it out.
From signals to decisions
The deeper problem is trust. Moving from static reporting to acting on AI-generated signals in real time — in environments where being wrong carries real consequences — isn't primarily a technology problem. It's about building the operational muscle to act on automated insight. To trust it enough to move, but stay awake enough that someone's still watching.
The organisations that get this right won't be the ones with the best models. They'll be the ones that figured out how to make AI-generated insight actionable inside their existing governance structures.
That's the problem I find myself working on. It's a good one.