Is the World Ready for Autonomous AI Agents, and are they Autonomous, or even Intelligent?
Craig Martin
Director | Strategic Innovation & Experimentation | Guiding clients to test and refine business decisions before investment and execution
If you’ve had a chance to explore OpenAI’s new Operator functionality, you’re probably as intrigued as I am about how these intelligent agents attempt to logically and autonomously solve problems for us. It’s fascinating—and in many ways echoes earlier technological promises we’ve heard before.
Many years ago, we experimented with ambitious tasks like: “Find me my top 10 customers, arrange customer visits, book flights, hotels, and taxis, and keep my boss and my wife updated about my travel plans.” At that time, Service-Oriented Architecture (SOA) was hailed as the solution to seamlessly automate these processes. Unfortunately, reality couldn’t quite live up to the hype.
Fast forward to today, and we find ourselves once again exploring the promise of intelligent automation, now through “Operators” and intelligent agents. Observing the OpenAI Operator, I saw first-hand the classic problem faced by agent automation: integrating smoothly with real-world applications. Currently, Operator agents have adopted a practical approach by effectively automating interactions via a web browser—this integration resembles Robotic Process Automation (RPA).
For instance, when I asked Operator to find and book the cheapest tickets for a Hans Zimmer concert in London, I watched—part amused, part frustrated—as it navigated through multiple different sites until it hit a roadblock. The gatekeeping systems are explicitly designed to stop automated bots. The irony wasn’t lost on me. Our modern internet landscape is littered with protective measures like CAPTCHAs, specifically designed to verify human identity. Yet we’re rapidly moving towards a future where sophisticated AI agents or avatars perform tasks on our behalf, blurring the line between human and bot interactions.
This introduces multiple layers of irony. We’ve previously witnessed disintermediation (remember the disappearance of travel agents?), effectively turning each of us into our own travel agents. Now we’re automating these tasks again, reinstating digital agents as intermediaries. But to realise this future smoothly, we’ll need to evolve gatekeeping systems from simplistic “human or bot” distinctions to more nuanced approaches acknowledging legitimate agent interactions.
This evolution raises critical governance and legal challenges. What’s the legal status of an AI agent making financial transactions or purchases on my behalf? If an agent makes a mistake or conducts unauthorised activities, who bears the responsibility? These challenges closely mirror debates surrounding autonomous vehicles—where liability for accidents remains an unresolved issue. Is responsibility held by the developer, the manufacturer, the vehicle owner, or the software itself? As always, regulatory and governance frameworks significantly lag behind technological advancements. Addressing these challenges demands urgent regulatory clarity, thoughtful policy-making, and robust governance mechanisms.
We may soon reach a tipping point where current gatekeeping measures—such as CAPTCHAs—become obsolete. Perhaps a new business opportunity lies in developing mechanisms for agents to securely identify and authenticate themselves through digital handshakes with updated, intelligent CAPTCHAs. Imagine websites confidently identifying specific named users via trusted agents. Monetising that authentication process through blockchain technologies could create genuinely exciting possibilities, where users willingly trade verified identities for discounts or perks—effectively “selling” their digital selves securely and transparently.
If you’ve experimented with agent-based offerings in the market, you will also have noticed how linear they are. Truth be told, it’s quite similar to the IFTTT setup that’s been around for a while, albeit more “intelligent” in determining its next step. However, true agent-based systems need to become more decentralised, autonomous, and self-learning.
This will leverage some of the more traditional branches of agent-based technology like MAS (Multi-Agent Systems), Complex Adaptive Systems (CAS), and Reinforcement Learning (RL). MAS isn’t new—it’s been around since at least the late 80s and early 90s. It’s a field of artificial intelligence that involves multiple autonomous agents interacting within a shared environment. These agents communicate, cooperate, negotiate, or compete to solve complex problems collectively, rather than individually. Recent advancements have led to the development of MCP—Model Context Protocol. MCP is essentially an emerging approach to standardising how AI models—particularly autonomous agents—manage and share context. It defines a structured way for agents to access, maintain, and synchronise contextual data, enabling better coordination, continuity, and more coherent interactions. Think of MCP as the “middleware for context”—allowing AI agents to remain synchronised, aware, and consistent as they independently and autonomously move across different environments, executing tasks, and interacting not only with each other but also with different internal and external environments.
Let’s make this practical by extending my current Operator use-case into a future-oriented version: Returning to my problem of automatically booking seats for a sold-out Hans Zimmer concert in London, an AI agent faces multiple challenges, some of which I experienced in the current iteration of OpenAI’s Operator:
Solving for this scenario requires sophisticated interaction and coordination among multiple autonomous agents—precisely where MAS, CAS, MCP, and RL become powerful allies.
1. Multi-Agent Systems (MAS):
MAS involves multiple autonomous agents working together. Here’s how it helps:
2. Complex Adaptive Systems (CAS):
Concert ticket availability, prices, and user behaviour form a complex adaptive system—continuously evolving and unpredictable:
3. Model Context Protocol (MCP):
MCP brings standardised, structured context management, significantly enhancing agents’ real-world effectiveness:
4. Reinforcement Learning (RL):
RL enables agents to improve strategies over time through continuous trial and error:
This comprehensive integration is where agent-based technologies really begin to deliver. From what I’ve observed, we still have some distance to cover before platforms offer this level of integrated multi-agent functionality. While certain pieces exist, no single integrated platform currently delivers it all. I reckon the first to master this integration solution will dominate the market.
It’s a fascinating new world we’re stepping into, full of potential—but also demanding careful consideration, strategic design, and well-defined governance. Are we ready for it?
Automation - AI Shepard - Strategy - Architecture - Delivery
1 周I am envious that you got your mittens on operator.