Skip to main content
  1. Blog/

Claude 3.5 Gets a Computer — Anthropic's 'Computer Use' and the Future of AI Agents

·1032 words·5 mins
Osmond van Hemert
Author
Osmond van Hemert
AI Models & Releases - This article is part of a series.
Part : This Article

The AI agent space has been buzzing with activity, and one development that keeps coming up in my conversations with other developers is Anthropic’s “computer use” capability for Claude. While it was initially announced in beta back in October, the real-world adoption and experimentation has been accelerating through early 2025. The idea is deceptively simple: give an AI model the ability to see a screen, move a mouse, and type on a keyboard — essentially letting it operate a computer the way a human would.

Having spent a few weeks experimenting with computer use in various automation scenarios, I wanted to share some thoughts on where this technology stands, where it’s genuinely useful, and where the hype outpaces reality.

How Computer Use Actually Works
#

The technical approach is straightforward in concept. Claude receives screenshots of the desktop, reasons about what it sees, and issues commands — mouse clicks at specific coordinates, keyboard input, scrolling. It’s essentially a vision-language model controlling a computer through the same interface a VNC user would.

The implementation typically involves running a containerized desktop environment (Anthropic provides Docker images for this), connecting Claude to it via their API, and defining the task you want accomplished. The model takes screenshots at each step, decides what to do next, and iterates until the task is complete or it gets stuck.

What’s remarkable is that this works without any application-specific integration. Claude doesn’t need an API for the application it’s controlling — it reads the screen and interacts with the UI. This means it can theoretically work with any desktop application, legacy system, or web interface, including ones that have no API at all.

Where It Actually Shines
#

After experimenting with various use cases, I’ve found computer use most compelling in a few specific scenarios:

Legacy system automation: Many organizations have critical business processes running through old desktop applications or web portals that were built before APIs were standard practice. Writing a traditional automation script for these systems is painful — you’re dealing with fragile screen scraping, COM automation, or reverse-engineering undocumented protocols. Computer use offers a higher-level abstraction: describe what you want done, and the AI figures out how to navigate the interface.

Testing workflows: Using computer use for end-to-end testing of complex web applications is intriguing. Rather than maintaining brittle Selenium scripts that break every time the UI changes, you can describe test scenarios in natural language. “Log in, navigate to the settings page, change the notification preferences, and verify the confirmation message.” The AI handles the implementation details.

Data entry and extraction: For tasks that involve copying data between systems — pulling information from one application and entering it into another — computer use eliminates the need for custom integration code. It’s not the most efficient approach, but for low-volume, high-variety tasks, it’s remarkably practical.

The Limitations Are Real
#

Let me temper the enthusiasm with some honest assessment of the current limitations.

Speed: Computer use is slow. Each interaction cycle involves taking a screenshot, sending it to the API, waiting for the model to reason about it, and executing the action. A task that a human could complete in 30 seconds might take several minutes. For high-volume automation, traditional scripted approaches are still far superior.

Reliability: The model makes mistakes. It misclicks, misreads text, gets confused by pop-ups or unexpected dialog boxes, and sometimes enters a loop of incorrect actions. In my testing, I’d estimate about a 70-80% success rate on moderately complex multi-step tasks. That’s impressive for an AI system but inadequate for production automation without human oversight.

Cost: Each step involves an API call with image input, which adds up quickly. A complex workflow might involve dozens of steps, each costing a few cents. For frequent automation tasks, the economics don’t currently favor computer use over traditional scripting.

Security implications: Giving an AI model control over a computer raises obvious security concerns. The model needs access to whatever the automated application can access, and a misguided action could have real consequences. Sandboxing and careful permission scoping are essential.

The Broader Agent Landscape
#

Computer use is part of a wider trend toward AI agents — systems that don’t just generate text but take actions in the real world. OpenAI has been pushing its own agent frameworks, Google’s Gemini has similar capabilities in development, and the open-source community has projects like Open Interpreter that offer comparable functionality.

What’s emerging is a spectrum of agent capabilities. At one end, you have tool-using models that call APIs and functions — relatively structured and predictable. At the other end, you have computer use, where the AI interacts with arbitrary interfaces through vision and motor control — flexible but less reliable.

The sweet spot for most production use cases is probably somewhere in the middle: agents that use structured tools (APIs, functions, databases) for core functionality, with computer use as a fallback for systems that don’t have programmatic interfaces.

My Take
#

I see computer use as a genuinely important capability, but one that’s currently better suited for prototyping, occasional automation, and handling edge cases than for production-scale operations. The technology will improve — models will get faster, more accurate, and cheaper — but the fundamental overhead of the screenshot-reason-act loop means it will likely remain slower than purpose-built integrations.

Where I’m most excited is the democratization of automation. Today, automating a workflow across multiple applications requires significant programming skill. Computer use lowers that barrier dramatically. A domain expert who can describe a process in plain language can now automate it, at least for personal productivity scenarios.

For us as developers, the implications are interesting. We should be thinking about how our applications will be used by AI agents — both through APIs (which should be the primary interface) and through UIs that agents can navigate. Accessible, well-structured interfaces aren’t just good for human users; they’re increasingly good for AI users too.

The agent era is coming, but it’s coming gradually, not as a sudden revolution. Computer use is one piece of that puzzle — a powerful but imperfect tool that’s worth understanding and experimenting with, even if it’s not ready to replace your CI/CD pipeline just yet.

AI Models & Releases - This article is part of a series.
Part : This Article