Jesse Reiss
CTO & Co-Founder
AI Agents are everywhere right now – in headlines, product demos, boardroom conversations. You understand why. The idea is intoxicating: software that doesn’t just answer your questions but can do the work for you.
And yet, depending on which headline you read, they’re either transforming business as we know it…or delivering absolutely zero return on investment.
Case in point: a recent MIT report found that a staggering 95% of generative AI pilot projects failed to deliver any measurable ROI at all. Ouch! But that stat (even if partially true) doesn’t mean the technology is broken. It means we’re still learning how – and when – to use it effectively. And in the meantime, the appetite for AI efficiency hasn’t abated. A recent PYMNTS survey found that 85% of executives now expect AI to meaningfully improve compliance operations.
So, which AI agents headline do we embrace? Do agents bring disappointment…or exponential potential?
The truth is: it depends on what you’re asking the AI to do.
When people ask “Do AI Agents work?” what they’re really asking is, “Can I trust an Agent to do this job?” And the answer – as with any employee, tool, or intern – depends entirely on the task you’ve assigned it.
That’s where we’ve found some thinking by Cedric Chin and Vaughn Tan to be extremely helpful. Tan, whose Harvard Business School doctorate focused on organizational behavior and sociology, writes about “meaning-making” tasks as being (still!) a uniquely human endeavor.
Between the two of them, we can create a conceptual distinction between two kinds of tasks:
What purpose does this distinction serve? Simple. By separating tasks into these two groups, we can identify which tasks to automate and which to leave to human judgment.
Some people might say, “does that distinction really matter?” And we would respond – if you’re working in financial compliance, we’d say it matters a lot.
It’s certainly mattered to us as we’ve built out Hummingbird’s AI features.
Let’s take a typical case investigation. Before an analyst can determine whether a behavior is suspicious, they need to:
This is all meaning-reducing work. The goal is clear – gather everything relevant so the analyst can make a decision. But it’s slow, repetitive, and error-prone. Perfect for an Agent.
Now let’s look at what comes next.
Once this initial phase is complete, more “meaning-making” work begins: investigators must interpret the data, apply policy and context, and ultimately make a judgment call. Do they escalate the review? Does the subject’s story check out? Is this risk real or merely a false positive?
This kind of work isn’t something we’d want to outsource to AI Agents (at least not today). And why?
Fortunately, this is exactly where human analysts shine! So if you hear us at Hummingbird say “human in the loop” AI compliance, this is what we’re talking about. This is the loop.
So, what does a successful Agent deployment look like? In our experience, it comes down to a few simple rules of thumb:
The conversation needs to shift. The question isn’t “Are Agents ready for prime time?” It’s “What kind of work are we asking them to do?
Because if you’re asking an Agent to replicate the judgment of a seasoned compliance analyst across a tangled web of policies and edge cases… you’re setting yourself up for disappointment (at least for now).
But if you’re asking it to gather and structure the raw material for that analyst so that they can focus on actual decision-making – well, you’re going to see your time (and results) improve exponentially.
That’s where the ROI is hiding. Not in vague promises of AI transformation, but in specific, well-chosen use cases that honor what the technology can do – and what only humans still should.
AI Agents are no more a cure-all than any other promising new technology.
Used intelligently, however, they’re an undeniable force multiplier.
At Hummingbird, we’re committed to deploying AI in ways that match both the technology’s strengths and the needs of compliance professionals. That means building Agents that speed up work without dumbing it down – through features that deliver clarity, not complexity.
If you’ve been navigating some of the questions we’ve been exploring in this essay and want to learn how to apply AI where it counts, we’d love to talk.
Let’s give Agents the right jobs – and watch what happens.
Subscribe to receive new content from Hummingbird