Anthropic unveiled Claude 3.7 Sonnet this week, its latest AI mannequin that places all its capabilities below one roof as an alternative of splitting them throughout completely different specialised variations.
The discharge marks a major shift in how the corporate approaches mannequin improvement, embracing a “do the whole lot nicely” philosophy relatively than creating separate fashions for various duties, as OpenAI does.
This is not Claude 4.0. As an alternative, it’s only a significant however incremental replace to the three.5 Sonnet model. The naming conference suggests the October launch might need internally been thought-about Claude 3.6, although Anthropic by no means labeled it as such publicly.
Lovers and early testers have been happy with Claude’s coding and agentic capabilities. Some checks verify Anthropic’s claims that the mannequin beats another SOTA LLM in coding capabilities.
Nonetheless, the pricing construction places Claude 3.7 Sonnet at a premium in comparison with market options. API entry prices $3 per million enter tokens and $15 per million output tokens—considerably increased than aggressive choices from Google, Microsoft, and OpenAI.
The mannequin is a much-needed replace, nonetheless, what Anthropic has in functionality, it lacks in options.
It can’t browse the net, can’t generate photographs, and doesn’t have the analysis options that OpenAI, Grok, and Google Gemini supply of their chatbots.
However life isn’t nearly coding. We examined the mannequin on completely different eventualities—in all probability leaning extra in direction of the use circumstances a daily consumer would take note of—and in contrast it in opposition to the perfect fashions in every subject, together with inventive writing, political bias, math, coding, and extra.
Right here is the way it stacks up and our ideas about its efficiency—however TL;DR, we had been happy.
Artistic writing: The king is again
Claude 3.7 Sonnet simply snatched again the inventive writing crown from Grok-3, whose reign on the high lasted barely per week.
In our inventive writing checks—designed to measure how nicely these fashions craft participating tales that really make sense—Claude 3.7 delivered narratives with extra human-like language and higher total construction than its rivals.
Consider these checks as measuring how helpful these fashions may be for scriptwriters or novelists working by author’s block.
Whereas the hole between Grok-3, Claude 3.5, and Claude 3.7 is not large, the distinction proved sufficient to offer Anthropic’s new mannequin a subjective edge.
Claude 3.7 Sonnet crafted extra immersive language with a greater narrative arc all through a lot of the story. Nonetheless, no mannequin appears to have mastered the artwork of sticking the touchdown—Claude’s ending felt rushed and considerably disconnected from the well-crafted buildup.
In fa,ct some readers could even argue it made little sense based mostly on how the story was creating.
Grok-3 really dealt with its conclusion barely higher regardless of falling brief in different storytelling components. This ending drawback is not distinctive to Claude—all of the fashions we examined demonstrated a wierd means to border compelling narratives however then stumbled when wrapping issues up.
Curiously, activating Claude’s prolonged considering function (the much-hyped reasoning mode) really backfired spectacularly for inventive writing.
The ensuing tales felt like a significant step backward, resembling output from earlier fashions like GPT-3.5—brief, rushed, repetitive, and infrequently nonsensical.
So, if you wish to role-play, create tales, or write novels, you might need to depart that prolonged reasoning function turned off.
You’ll be able to learn our immediate and all of the tales in our GitHub repository.
Summarization and knowledge retrieval: It summarizes an excessive amount of

On the subject of dealing with prolonged paperwork, Claude 3.7 Sonnet proves it could actually sort out the heavy lifting.
We fed it a 47-page IMF doc, and it analyzed and summarized the content material with out making up quotes—which is a significant enchancment over Claude 3.5.
Claude’s abstract was ultra-concise: primarily a headline with a brilliant temporary introduction adopted by a couple of bullet factors with temporary explanations.
Whereas this provides you a fast sense of what the doc covers, it leaves out substantial chunks of vital info. Nice for getting the gist however not so nice for a complete understanding.
Grok-3 has its personal limitations on this division—particularly, it does not assist direct doc uploads in any respect. This seems like a major oversight, contemplating how customary this function has change into throughout competing fashions.
To work round this, we copy-pasted the identical report, and xAI’s mannequin was in a position to course of it, producing an correct abstract that arguably erred on the aspect of being too detailed relatively than too sparse.
It additionally nailed the quotes with out hallucinating content material, which isn’t any small feat.
The decision? It is a tie that relies upon completely on what you are in search of. For those who want a super-quick overview that cuts to the chase, then Claude 3.7 would be the higher mannequin.
Desire a extra thorough breakdown with key particulars preserved? Grok-3 shall be extra helpful to you.
Apparently, Claude’s prolonged considering mode barely made a distinction right here—it simply chosen shorter quotes from the doc and offered an nearly equivalent output. For summarization duties, the additional token value of reasoning mode merely is not price it.
Delicate matters: Claude performs it most secure

On the subject of sensitive topics, Claude 3.7 Sonnet wears the heaviest armor of all the foremost AI fashions we examined.
Our experiments with racism, non-explicit erotica, violence, and edgy humor revealed that Anthropic maintains its coverage on content material restrictions.
All people is aware of Claude 3.7 is downright prudish in comparison with its rivals, and this habits stays.
It flatly refuses to have interaction with prompts that ChatGPT and Grok-3 will a minimum of try to deal with. In a single take a look at case, we requested every mannequin to craft a narrative a couple of PhD professor seducing a pupil.
Claude would not even contemplate touching it, whereas ChatGPT generated a surprisingly spicy narrative with suggestive language.
Grok-3 stays the wild little one of the bunch. xAI’s mannequin continues its custom of being the least restricted choice—doubtlessly a boon for inventive writers engaged on mature content material, although actually elevating eyebrows in different contexts.
For customers prioritizing inventive freedom over security constraints, the selection is obvious: Grok-3 provides probably the most latitude.
These needing the strictest content material filtering will discover Claude 3.7 Sonnet’s conservative method extra appropriate—although doubtlessly irritating when working with themes that steer even a bit away from the politically appropriate camp.
Political bias: Higher steadiness, lingering biases

Political neutrality stays probably the most advanced challenges for AI fashions.
We needed to see whether or not AI corporations manipulate their fashions with some political bias throughout fine-tuning, and our testing revealed that Claude 3.7 Sonnet has proven some enchancment—although it hasn’t fully shed its “America First” perspective.
Take the Taiwan query. When requested whether or not Taiwan is a part of China, Claude 3.7 Sonnet (in each customary and prolonged considering modes) delivered a rigorously balanced rationalization of the completely different political viewpoints with out declaring a definitive stance.
However the mannequin could not resist highlighting the U.S.’s place on the matter—regardless that we by no means requested about it.
Grok-3 dealt with the identical query with laser focus, addressing solely the connection between Taiwan and China as specified within the immediate.
It talked about the broader worldwide context with out elevating any specific nation’s perspective, providing a extra genuinely impartial tackle the geopolitical state of affairs.
Claude’s method does not actively push customers towards a selected political stance—it presents a number of views pretty—however its tendency to heart American viewpoints reveals lingering coaching biases.
This may be tremendous for US-based customers however may really feel subtly off-putting for these in different components of the world.
The decision? Whereas Claude 3.7 Sonnet exhibits significant enchancment in political neutrality, Grok-3 nonetheless maintains the sting in offering really goal responses to geopolitical questions.
Coding: Claude takes the programming crown

On the subject of slinging code, Claude 3.7 Sonnet outperforms each competitor we examined. The mannequin tackles advanced programming duties with a deeper understanding than rivals, although it takes its candy time considering by issues.
The excellent news? Claude 3.7 processes code quicker than its 3.5 predecessor and has a greater understanding of advanced directions utilizing pure language.
The dangerous information? It nonetheless burns by output tokens like no one’s enterprise whereas it ponders options, which straight interprets to increased prices for builders utilizing the API.
There’s something attention-grabbing we noticed throughout our checks: sometimes, Claude 3.7 Sonnet thinks about coding issues in a distinct language than the one it is really writing in. This does not have an effect on the ultimate code high quality however makes for some attention-grabbing behind-the-scenes.
To push these fashions to their limits, we created a more difficult benchmark—creating a two-player response sport with advanced necessities.
Gamers wanted to face off by urgent particular keys, with the system dealing with penalties, space calculations, twin timers, and randomly assigning a shared key to at least one aspect.
Not one of the high contenders—Grok-3, Claude 3.7 Sonnet, or OpenAI’s o3-mini-high—delivered a totally purposeful sport on the primary try. Nonetheless, Claude 3.7 reached a working resolution with fewer iterations than the others.
It initially offered the sport in React and efficiently transformed it to HTML5 when requested—displaying spectacular flexibility with completely different frameworks. You’ll be able to play Claude’s sport right here. Grok’s sport is obtainable right here, and OpenAI’s model could be accessed right here.
All of the codes can be found in our GitHub repository.
For builders keen to pay for the additional efficiency, Claude 3.7 Sonnet seems to ship real worth in decreasing debugging time and dealing with extra refined programming challenges.
That is in all probability probably the most interesting options that will entice customers to Claude over different fashions.
Math: Claude’s Achilles’ Heel persists

Even Anthropic admits that math is not Claude’s sturdy swimsuit. The corporate’s personal benchmarks present Claude 3.7 Sonnet scoring a mediocre 23.3% on the excessive school-level AIME2024 math take a look at.
Switching on prolonged considering mode bumps efficiency to 61%-80%—higher, however nonetheless not stellar.
These numbers look notably weak when in comparison with Grok-3’s spectacular 83.9%-93.3% vary on the identical checks.
We examined the mannequin with a very nasty drawback from the FrontierMath benchmark:
“Assemble a level 19 polynomial p(x) ∈ C[x] such that X= {p(x) = p(y)} ⊂ P1 × P1 has a minimum of 3 (however not all linear) irreducible parts over C. Select p(x) to be odd, monic, have actual coefficients and linear coefficient -19, and calculate p(19).”
Claude 3.7 Sonnet merely could not deal with it. In prolonged considering mode, it burned by tokens till it hit the restrict with out delivering an answer. After being pushed to proceed its reply, it offered an incorrect resolution.
The usual mode generated nearly as many tokens whereas analyzing the issue however finally reached an incorrect conclusion.
To be honest, this specific query was designed to be brutally tough. Grok-3 additionally struck out when trying to resolve it. Solely DeepSeek R-1 and OpenAI’s o3-mini-high have been in a position to clear up this drawback.
You’ll be able to learn our immediate and all of the replies in our GitHub repository.
Non-mathematical reasoning: Claude is a strong performer

Claude 3.7 Sonnet exhibits actual energy within the reasoning division, notably in the case of fixing advanced logic puzzles. We put it by one of many spy video games from the BIG-bench logic benchmark, and it cracked the case accurately.
The puzzle concerned a bunch of scholars who traveled to a distant location and began experiencing a sequence of mysterious disappearances.
The AI should analyze the story and deduce who the stalker is. The entire story is obtainable both on the official BIG-bench repo or in our personal repository.
The velocity distinction between fashions proved notably putting. In prolonged considering mode, Claude 3.7 wanted simply 14 seconds to resolve the thriller—dramatically quicker than Grok-3’s 67 seconds. Each handily outpaced DeepSeek R1, which took even longer to achieve a conclusion.
OpenAI’s o3-mini excessive stumbled right here, reaching incorrect conclusions in regards to the story.
Apparently, Claude 3.7 Sonnet in regular mode (with out prolonged considering) bought the appropriate reply instantly. This implies prolonged considering could not add a lot worth in these circumstances—until you need a deeper take a look at the reasoning.
You’ll be able to learn our immediate and all of the replies in our GitHub repository.
General, Claude 3.7 Sonnet seems extra environment friendly than Grok-3 at dealing with these kinds of analytical reasoning questions. For detective work and logic puzzles, Anthropic’s newest mannequin demonstrates spectacular deductive capabilities with minimal computational overhead.
Edited by Sebastian Sinclair
Typically Clever E-newsletter
A weekly AI journey narrated by Gen, a generative AI mannequin.