Skip to main content
  1. Blog/

Claude 3.5 Sonnet — Anthropic Raises the Bar for Coding AI

·1030 words·5 mins
Osmond van Hemert
Author
Osmond van Hemert
AI Models & Releases - This article is part of a series.
Part : This Article

The AI model landscape just shifted again. Today, Anthropic released Claude 3.5 Sonnet, and the benchmarks are turning heads. The mid-tier model in their lineup is outperforming GPT-4o on most coding benchmarks while running at twice the speed and one-fifth the cost of their previous top model, Claude 3 Opus. For those of us who use LLMs as daily development tools, this isn’t just another incremental update — it’s a meaningful shift in what’s available.

The Numbers That Matter
#

Let’s start with the benchmarks that are most relevant to developers. On HumanEval, the standard coding benchmark, Claude 3.5 Sonnet scores 92.0% — surpassing GPT-4o’s 90.2%. On the graduate-level reasoning benchmark (GPQA), it hits 59.4%, compared to GPT-4o’s 53.6%. But the benchmark that caught my eye is the internal “agentic coding” evaluation Anthropic shared, where Claude 3.5 Sonnet solved 64% of problems compared to Claude 3 Opus at 38%.

Benchmarks are always somewhat synthetic, and I’ve been in this field long enough to know that real-world performance often diverges from leaderboard scores. But the early anecdotal evidence from developers aligns with these numbers. Multi-file refactoring tasks, understanding complex codebases, and generating tests all seem measurably improved.

The speed improvement is equally important. Sonnet operates at roughly twice the token throughput of Opus, which means faster completions in interactive coding sessions. When you’re using an AI assistant in your IDE, the difference between a 3-second and a 6-second response significantly affects your flow state. I’ve been feeling this friction with Opus for a while — capable but slow enough to break concentration.

Artifacts: Rethinking the Chat Interface
#

Alongside the model release, Anthropic introduced Artifacts — a new feature in the Claude interface that renders generated content in a dedicated panel alongside the conversation. When you ask Claude to write code, create a document, or build a simple application, the output appears in an interactive preview that you can iterate on.

This might sound like a UI detail, but it represents a shift in how we interact with AI coding assistants. Instead of copying code from a chat window into your editor, you get a live preview that you can modify and refine within the conversation. For quick prototyping, creating utility scripts, or building proof-of-concept applications, this workflow is remarkably efficient.

I spent an hour today building a data visualization dashboard through Artifacts — describing what I wanted, seeing it rendered immediately, and iterating on the design through conversation. The entire process from concept to working prototype took about 20 minutes. That same task would have taken me half a day with traditional development, including the inevitable fiddling with chart library documentation.

The Competitive Landscape Shifts
#

What makes this release strategically significant is the pricing model. Claude 3.5 Sonnet is priced at $3 per million input tokens and $15 per million output tokens — the same as Claude 3 Sonnet, their previous mid-tier model. You’re getting performance that exceeds the previous flagship at the mid-tier price point. Anthropic is effectively making top-tier AI coding assistance significantly more accessible.

This creates real competitive pressure on OpenAI. GPT-4o is priced similarly, but if Claude 3.5 Sonnet consistently outperforms it on coding tasks, developers building AI-powered tools will start defaulting to Anthropic’s API. The downstream effects on products like Cursor, Cody, and other AI coding tools could be significant — many of these tools offer model selection, and users will gravitate toward better results.

Google’s Gemini 1.5 Pro is also in this competitive mix, and its million-token context window gives it unique advantages for large codebase analysis. But on pure code generation quality, Claude 3.5 Sonnet appears to have the edge right now.

What This Means for Developer Workflows
#

I’ve been using AI coding assistants since the early days of Copilot, and I’ve watched the capability curve closely. We’re reaching a point where these tools are genuinely useful for substantive programming tasks, not just autocomplete. Specific areas where I’m seeing real productivity gains with this class of model:

Code review and refactoring. Point Claude 3.5 Sonnet at a messy function and ask it to refactor with specific constraints (maintain the API, improve error handling, add typing). The suggestions are increasingly production-quality.

Test generation. Describe the edge cases you’re worried about, and the model generates comprehensive test suites that actually catch bugs. This has been a weak point of earlier models, but the improvement is notable.

Documentation. Generating accurate docstrings, README files, and API documentation from code. The model understands intent well enough that the documentation reads naturally, not like machine-generated boilerplate.

Debugging complex issues. Paste in a stack trace, relevant code, and a description of the expected behavior. The diagnostic reasoning is significantly better than previous models — it considers multiple hypotheses and asks clarifying questions.

The 200K context window (same as Claude 3 Opus) means you can feed substantial portions of a codebase into a single conversation. For understanding how a change propagates through a system, this is invaluable.

My Take
#

The AI coding assistant space is maturing faster than I expected. Six months ago, I was treating these tools as useful-but-unreliable helpers that needed constant verification. With Claude 3.5 Sonnet, I’m starting to trust the output enough to use it for more critical tasks — still with review, but with less overhead.

What I find most interesting is the business model dynamic. Anthropic is releasing what is arguably the best coding AI available, at mid-tier pricing. This pushes the entire market toward making top-tier AI accessible to individual developers and small teams, not just enterprises with large API budgets.

The real question is sustainability. Training and serving these models is extraordinarily expensive, and the pricing suggests companies are prioritizing market share over margins. That race benefits developers in the short term, but I wonder about the long-term equilibrium.

For now, though, if you’re a developer not yet using AI assistance in your workflow, Claude 3.5 Sonnet might be the model that changes your mind. The quality-to-cost ratio has crossed a threshold that makes it practical for daily use, not just occasional experimentation. I’ve updated my default model in every tool that offers the choice. The improvement is that clear.

AI Models & Releases - This article is part of a series.
Part : This Article