Anthropic has launched Claude Opus 4.1, an improve to its flagship mannequin that’s stated to ship higher efficiency in coding, reasoning, and autonomous job dealing with.
The brand new mannequin is obtainable now to Claude Professional customers, Claude Code subscribers, and builders utilizing the API, Amazon Bedrock, or Google Cloud’s Vertex AI.
Efficiency Beneficial properties
Claude Opus 4.1 scores 74.5% on SWE-bench Verified, a benchmark for real-world coding issues, and is positioned as a drop-in alternative for Opus 4.
The mannequin exhibits notable enhancements in multi-file code refactoring and debugging, significantly in giant codebases. In accordance with GitHub and enterprise suggestions cited by Anthropic, it outperforms Opus 4 in most coding duties.
Rakuten’s engineering staff studies that Claude 4.1 exactly identifies code fixes with out introducing pointless modifications. Windsurf, a developer platform, measured a one customary deviation efficiency acquire in comparison with Opus 4, akin to the leap from Claude Sonnet 3.7 to Sonnet 4.
Expanded Use Circumstances
Anthropic describes Claude 4.1 as a hybrid reasoning mannequin designed to deal with each prompt outputs and prolonged pondering. Builders can fine-tune “pondering budgets” through the API to stability price and efficiency.
Key use circumstances embrace:
- AI Brokers: Robust outcomes on TAU-bench and long-horizon duties make the mannequin appropriate for autonomous workflows and enterprise automation.
- Superior Coding: With assist for 32,000 output tokens, Claude 4.1 handles advanced refactoring and multi-step era whereas adapting to coding model and context.
- Information Evaluation: The mannequin can synthesize insights from giant volumes of structured and unstructured information, similar to patent filings and analysis papers.
- Content material Technology: Claude 4.1 generates extra pure writing and richer prose than earlier variations, with higher construction and tone.
Security Enhancements
Claude 4.1 continues to function below Anthropic’s AI Security Stage 3 customary. Though the improve is taken into account incremental, the corporate voluntarily ran security evaluations to make sure efficiency stayed inside acceptable danger boundaries.
- Harmlessness: The mannequin refused policy-violating requests 98.76% of the time, up from 97.27% with Opus 4.
- Over-refusal: On benign requests, the refusal fee stays low at 0.08%.
- Bias and Youngster Security: Evaluations discovered no important regression in political bias, discriminatory conduct, or baby security responses.
Anthropic additionally examined the mannequin’s resistance to immediate injection and agent misuse. Outcomes confirmed comparable or improved conduct over Opus 4, with extra coaching and safeguards in place to mitigate edge circumstances.
Wanting Forward
Anthropic says bigger upgrades are on the horizon, with Claude 4.1 positioned as a stability-focused launch forward of future leaps.
For groups already utilizing Claude Opus 4, the improve path is seamless, with no modifications to API construction or pricing.
Featured Picture: Ahyan Inventory Studios/Shutterstock