A recent study from Zurich University of Applied Sciences by Pascal J. Sager, Benjamin Meyer, Peng Yan, Rebekka von Wartburg-Kottler, Layan Etaiwi, Aref Enayati, Gabriel Nobel, Ahmed Abdulkadir, Benjamin F. Grewe, and Thilo Stadelmann reveals that AI agents have officially outgrown their chatbot phase.
AI agents are running the show, clicking, scrolling, and typing their way through workflows with eerie precision. These instruction-based computer control agents (CCAs) can execute commands, interacting with digital environments like seasoned human operators. But as they edge closer to full autonomy, one thing becomes clear: the more power we give them, the harder it becomes to keep them in check.
How AI agents are learning to use computers like you
Traditional automation tools are glorified macros—repetitive, rigid, and clueless outside their scripted paths. CCAs, on the other hand, are built to improvise. They don’t just follow instructions; they observe, interpret, and act based on what they “see” on a screen, thanks to vision-language models (VLMs) and large language models (LLMs). This allows them to:
- Read screens like a human, identifying text, buttons, and input fields without predefined coordinates.
- Execute multi-step tasks, like opening an email, copying data, pasting it into a spreadsheet, and hitting send—all without direct supervision.
- Understand natural language instructions, removing the need for users to learn complex automation scripts.
- Adapt to changing interfaces, making them significantly more flexible than rule-based automation tools.
Tell a CCA to “find today’s top sales leads and email them a follow-up,” and it moves through apps, extracts relevant data, composes an email, and sends it, just like a human assistant. Unlike old-school RPA (Robotic Process Automation) that falls apart when a UI changes, CCAs can adjust in real time, identifying visual elements and making decisions on the fly.
The next frontier? Integration with cloud-based knowledge repositories and autonomous decision-making. The more these agents learn, the more sophisticated their capabilities become—raising questions about just how much trust we should place in them.
How large language models are transforming peer review
The benefits: Productivity, accessibility, and automation
There’s no denying that CCAs come with serious advantages:
- Productivity on steroids: Tedious, time-consuming tasks vanish, allowing workers to focus on higher-value decisions rather than clicking through dashboards.
- Accessibility revolution: People with disabilities can interact with technology more seamlessly through AI-powered navigation and task automation.
- Enterprise-wide scalability: Businesses can automate entire workflows without hiring an army of IT specialists to build custom solutions.
- System-wide integration: CCAs work across different platforms and applications, ensuring seamless digital interactions.
- Always-on efficiency: Unlike human workers, these agents don’t get tired, distracted, or take lunch breaks.
The risks: Privacy, security, and trust
For every productivity win, there’s an equal and opposite security nightmare lurking in the background. Giving AI control over user interfaces isn’t just automation—it’s granting an unblinking machine access to sensitive workflows, financial transactions, and private data. And that’s where things get complicated.
CCAs operate by “watching” screens and analyzing text. Who ensures that sensitive information isn’t being misused or logged? Who’s keeping AI-driven keystrokes in check?
If an AI agent can log into your banking app and transfer money with a single command, what happens if it’s hacked? We’re handing over the digital keys to the kingdom with few safeguards. If a CCA makes a catastrophic error—deletes the wrong file, sends the wrong email, or approves a disastrous transaction—who’s responsible? Humans can be fired, fined, or trained. AI? Not so much.
And, if a malicious actor hijacks a CCA, they don’t just get access—they get a tireless, automated accomplice capable of wreaking havoc at scale. Lawmakers are scrambling to keep up, but there’s no playbook for AI-driven digital assistants making high-stakes decisions in real-time.
What comes next?
Businesses are moving cautiously, trying to balance the undeniable efficiency gains with the looming risks. Some companies are enforcing “human-in-the-loop” models, where AI agents handle execution but require manual approval for critical actions. Others are investing in AI governance policies to create safeguards before these agents become standard in enterprise operations.
What’s certain is that CCAs aren’t a passing trend—they’re the next phase of AI evolution, quietly embedding themselves into workflows and interfaces everywhere. As they grow more capable, the debate won’t be about whether we should use them, but how we can possibly control them.
Images: Kerem Gülen/Midjourney