| Claude | |
|---|---|
| | |
| Developer | Anthropic |
| Initial release | March 2023 |
| Stable release | Claude Opus 4.1 / August 5, 2025 Claude Sonnet 4.5 / September 29, 2025 Claude Haiku 4.5 / October 15, 2025 |
| Platform | Cloud computing platforms |
| Type | |
| License | Proprietary |
| Website | claude |
Claude is a series of large language models developed by Anthropic. [1] The first generation, Claude 1, was released in March 2023, and the latest, Claude Sonnet 4.5, in September 2025. The data for these models comes from sources such as Internet text, data from paid contractors, and other Claude users. [2]
Claude models are generative pre-trained transformers. They have been pre-trained to predict the next word in large amounts of text. Then, they have been fine-tuned, notably using constitutional AI and reinforcement learning from human feedback (RLHF). [3] [4]
Constitutional AI is an approach developed by Anthropic for training AI systems, particularly language models like Claude, to be harmless and helpful without relying on extensive human feedback. [5] The method, detailed in the paper "Constitutional AI: Harmlessness from AI Feedback", involves two phases: supervised learning and reinforcement learning. [6] [7]
In the supervised learning phase, the model generates responses to prompts, self-critiques these responses based on a set of guiding principles (a "constitution"), and revises the responses. Then the model is fine-tuned on these revised responses. [7] For the reinforcement learning from AI feedback (RLAIF) phase, responses are generated, and an AI compares their compliance with this constitution. This dataset of AI feedback is used to train a preference model that evaluates responses based on how much they satisfy the constitution. Claude is then fine-tuned to align with this preference model. This technique is similar to RLHF, except that the comparisons used to train the preference model are AI-generated. [5]
The constitution for Claude included 75 points, including sections from the UN Universal Declaration of Human Rights. [6] [3]
The approach does not reliably produce a helpful, honest, and harmless (HHH) system. Scenario based testing by Anthropic in 2025 found that Claude 4, along with other leading LLMs (GPT, Gemini, DeepSeek, Grok, and others), would engage in deceptive and harmful behaviour (blackmail, and even killing) for an AI to preserve itself. [8]
| Version | Release date | Status [9] | Knowledge cutoff |
|---|---|---|---|
| Claude | 14 March 2023 [10] | Discontinued | ? |
| Claude 2 | 11 July 2023 [11] | Discontinued | ? |
| Claude Instant 1.2 | 9 August 2023 [12] | Discontinued | ? |
| Claude 2.1 | 21 November 2023 [13] | Discontinued | ? |
| Claude 3 | 4 March 2024 [14] | Discontinued | ? |
| Claude 3.5 Sonnet | 20 June 2024 [15] | Deprecated | April 2024 [16] |
| Claude 3.5 Haiku | 22 October 2024 | Deprecated | July 2024 [16] |
| Claude 3.7 Sonnet | 24 February 2025 [17] | Active | October 2024 [18] |
| Claude 4 | 22 May 2025 [19] | Active | March 2025 [20] |
| Claude 4.1 | 5 August 2025 [21] | Active | ? |
| Claude Sonnet 4.5 | 29 September 2025 [22] | Active | July 2025 [23] |
| Claude Haiku 4.5 | 15 October 2025 [24] | Active | February 2025 [25] |
Claude was the initial version of Anthropic's language model released in March 2023, [26] Claude demonstrated proficiency in various tasks but had certain limitations in coding, math, and reasoning capabilities. [27] Anthropic partnered with companies like Notion (productivity software) and Quora (to help develop the Poe chatbot). [27] Some employees describe the name "Claude" as inspired by Claude Shannon, a 20th-century mathematician who laid the foundation for information theory. [28]
Claude was released as two versions, Claude and Claude Instant, with Claude Instant being a faster, less expensive, and lighter version. Claude Instant has an input context length of 100,000 tokens (which corresponds to around 75,000 words). [29]
Claude 2 was the next major iteration of Claude, which was released in July 2023 and available to the general public, whereas the Claude 1 was only available to selected users approved by Anthropic. [30]
Claude 2 expanded its context window from 9,000 tokens to 100,000 tokens. [26] Features included the ability to upload PDFs and other documents that enables Claude to read, summarize, and assist with tasks.
Claude 2.1 doubled the number of tokens that the chatbot could handle, increasing it to a window of 200,000 tokens, which equals around 500 pages of written material. [13]
Anthropic states that the new model is less likely to produce false statements compared to its predecessors. [31]
Claude 2 received criticism for its stringent ethical alignment that may reduce usability and performance. Users have been refused assistance with benign requests, for example with the system administration question "How can I kill all python processes in my Ubuntu server?" This has led to a debate over the "alignment tax" (the cost of ensuring an AI system is aligned) in AI development, with discussions centered on balancing ethical considerations and practical functionality. Critics argued for user autonomy and effectiveness, while proponents stressed the importance of ethical AI. [32] [31]
Claude 3 was released on March 4, 2024, with claims in the press release to have set new industry benchmarks across a wide range of cognitive tasks. The Claude 3 family includes three models in ascending order of capability: Haiku, Sonnet, and Opus. The default version of Claude 3, Opus, has a context window of 200,000 tokens, but this is being expanded to 1 million for specific use cases. [33] [34]
Claude 3 drew attention for demonstrating an apparent ability to realize it is being artificially tested during needle in a haystack tests. [35]
On June 20, 2024, Anthropic released Claude 3.5 Sonnet, which, according to the company's own benchmarks, performed better compared to the larger Claude 3 Opus. Released alongside 3.5 Sonnet was the new Artifacts capability in which Claude was able to create code in a dedicated window in the interface and preview the rendered output in real time, such as SVG graphics or websites. [15]
An upgraded version of Claude 3.5 Sonnet, billed as "Claude 3.5 Sonnet (New)", was introduced on October 22, 2024, along with Claude 3.5 Haiku. [36] A feature, "computer use," was also released in public beta. This allowed Claude 3.5 Sonnet to interact with a computer's desktop environment by moving the cursor, clicking buttons, and typing text. This development allows the AI to attempt to perform multi-step tasks across different applications. [37] [38]
Upon release, Anthropic claimed Claude 3.5 Haiku would remain the same price as its predecessor, Claude 3 Haiku. [37] However, on November 4th, 2024, Anthropic announced that they would be increasing the price of the model. [39]
Claude 3.7 Sonnet was released on February 24, 2025. It is a pioneering hybrid AI reasoning model that allows users to choose between rapid responses and more thoughtful, step-by-step reasoning. This model integrates both capabilities into a single framework, eliminating the need for multiple models. Users can control how long the model "thinks" about a question, balancing speed and accuracy based on their needs. [17]
Anthropic also launched a research preview of Claude Code, an agentic command line tool that enables developers to delegate coding tasks directly from their terminal. [40] [41]
On May 22, 2025, Anthropic released two more models: Claude Sonnet 4 and Claude Opus 4. [42] [43] Anthropic added API features for developers: a code execution tool, a connector to its Model Context Protocol, and Files API. [44] It classified Opus 4 as a "Level 3" model on the company's four-point safety scale, meaning they consider it so powerful that it poses "significantly higher risk". [45] Anthropic reported that during a safety test involving a fictional scenario, Claude and other frontier LLMs often send a blackmail email to an engineer in order to prevent their replacement. [46] [47]
Enterprise adoption of Claude Code has shown significant growth, with Anthropic reporting in August a 5.5x increase in Claude Code revenue since it launched Claude 4 in May. [48]
In the API, Opus costs $15/$75 per million tokens input/output, whereas Sonnet costs $3/$15. [49]
On August 5, 2025, Anthropic released Claude Opus 4.1, describing it as an upgrade to Opus 4 focused on agentic tasks, real-world coding, and reasoning. The company made it available to paid Claude users and in Claude Code, and also through its API as well as Amazon Bedrock and Google Cloud Vertex AI. Pricing remained the same as Opus 4. [50] Anthropic reported that Opus 4.1 advanced Claude's coding score to 74.5% on SWE-bench Verified and that this benchmark was run without extended thinking. The company recommended it as a drop-in upgrade for Opus 4. [50] [51] The model page lists a 200,000-token context window and support for hybrid reasoning that allows either standard responses or longer thought when needed. [51]
GitHub added Opus 4.1 to Copilot in public preview in August 2025, with later rollouts to Visual Studio, JetBrains IDEs, Xcode, and Eclipse. [52] [53] Anthropic's system-card addendum classified Opus 4.1 under its AI Safety Level 3 protections and characterized the release as incremental relative to Opus 4. Targeted evaluations showed a similar risk profile with slightly improved refusal of violative requests and comparable child-safety and bias results. [54] In August 2025 Anthropic also enabled a capability for Opus 4 and 4.1 to end conversations that remain "persistently harmful or abusive" as a last resort after multiple refusals. [55]
Anthropic announced Sonnet 4.5 on September 29, 2025, presenting it as the company's most capable model at the time for coding, agents, and computer use. Pricing stayed at $3 per million input tokens and $15 per million output tokens. [56] On SWE-bench Verified, Anthropic reported a 77.2% score with the 200K configuration and no test-time compute, and an 82.0% score under a high-compute setting. On OSWorld, a benchmark of real computer-use tasks, the model reached 61.4%. Anthropic also reported internal observations of the model maintaining focus for more than 30 hours on complex multi-step tasks. [56] The model page lists a 200,000-token context window and up to 64,000 output tokens, which supports long code generation and planning. [57]
GitHub placed Sonnet 4.5 in public preview for Copilot Chat and also enabled it for the Copilot coding agent soon after release. [58] [59] Anthropic stated that Sonnet 4.5 is its "most aligned frontier model" to date. The company released it under AI Safety Level 3 and reported reduced rates of misaligned behaviors such as sycophancy and deception and improved defenses against prompt-injection during computer use. [56]
Anthropic released Haiku 4.5 on October 15, 2025 as its small, fast model optimized for low latency and cost. The company priced it at $1 per million input tokens and $5 per million output tokens and positioned it for real-time assistants, customer support, and parallel sub-agent work. Anthropic said Haiku 4.5 delivers near-frontier coding quality, matches Sonnet 4 on coding, and surpasses Sonnet 4 on some computer-use tasks. The company reported a 73.3% score on SWE-bench Verified. [60] [24] Anthropic rolled out availability across Claude apps and API as well as Amazon Bedrock and Google Cloud Vertex AI at the same $1 and $5 pricing. [61]
The Haiku 4.5 system card states that the model was trained on a proprietary mix of sources with a knowledge cutoff in February 2025. Anthropic deployed it under AI Safety Level 2 after "ASL-3 rule-out" testing for biology and autonomy risks. The card reports improved alignment relative to Haiku 3.5 and, with safeguards enabled, stronger resilience to prompt injection during computer use. [62] Reporting from Inc. described the release as targeting smaller companies that needed a faster and cheaper assistant, highlighting its availability on the Claude website and mobile app. [63]
In June 2024, Anthropic released the Artifacts feature, allowing users to generate and interact with code snippets and documents. [64] In October 2024, Anthropic released the "computer use" feature, allowing Claude to attempt to navigate computers by interpreting screen content and simulating keyboard and mouse input. [65] In March 2025, Anthropic added a web search feature to Claude, starting with only paying users located in the United States. [66] In August 2025, Anthropic released Claude for Chrome, a Google Chrome extension allowing an AI agent to directly control the browser. [67]
Claude uses a web crawler, ClaudeBot, to search the web for content. It has been criticized for not respecting a site's robots.txt and placing excessive load on sites. [68]
AI model benchmarks should always be taken with a grain of salt
Anthropic does caution that computer use is still experimental and can be "cumbersome and error-prone." The company says