option
Home
List of Al models
DeepSeek-R1
Model parameter quantity
671B
Model parameter quantity
Affiliated organization
DeepSeek
Affiliated organization
Open Source
License Type
Release time
January 20, 2025
Release time

Model Introduction
DeepSeek-R1 extensively utilized reinforcement learning techniques during the post-training phase, significantly enhancing the model's reasoning capabilities with only a minimal amount of annotated data. On tasks involving mathematics, coding, and natural language inference, its performance is on par with the official release of OpenAI's o1.
Swipe left and right to view more
Language comprehension ability Language comprehension ability
Language comprehension ability
Capable of understanding complex contexts and generating logically coherent sentences, though occasionally off in tone control.
7.5
Knowledge coverage scope Knowledge coverage scope
Knowledge coverage scope
Covers more than 200 specialized fields, integrating the latest research findings and cross-cultural knowledge in real time.
9.0
Reasoning ability Reasoning ability
Reasoning ability
Can perform logical reasoning with more than three steps, though efficiency drops when handling nonlinear relationships.
8.5
Related model
DeepSeek-V3-0324 DeepSeek-V3 outperforms other open-source models such as Qwen2.5-72B and Llama-3.1-405B in multiple evaluations and matches the performance of top-tier closed-source models like GPT-4 and Claude-3.5-Sonnet.
DeepSeek-R1-0528 The latest version of Deepseek R1.
DeepSeek-V2-Chat-0628 DeepSeek-V2 is a strong Mixture-of-Experts (MoE) language model characterized by economical training and efficient inference. It comprises 236B total parameters, of which 21B are activated for each token. Compared with DeepSeek 67B, DeepSeek-V2 achieves stronger performance, and meanwhile saves 42.5% of training costs, reduces the KV cache by 93.3%, and boosts the maximum generation throughput to 5.76 times.
DeepSeek-V2.5 DeepSeek-V2.5 is an upgraded version that combines DeepSeek-V2-Chat and DeepSeek-Coder-V2-Instruct. The new model integrates the general and coding abilities of the two previous versions.
DeepSeek-V3-0324 DeepSeek-V3 outperforms other open-source models such as Qwen2.5-72B and Llama-3.1-405B in multiple evaluations and matches the performance of top-tier closed-source models like GPT-4 and Claude-3.5-Sonnet.
Relevant documents
AI Ad Scaling Revolution: Supercharge Creativity by 10X in 2025 The digital advertising landscape continues its rapid evolution, making innovation imperative for competitive success. As we approach 2025, the fusion of artificial intelligence and creative marketing presents groundbreaking opportunities to revoluti
AI Recruitment Systems Expose Hidden Biases Impacting Hiring Decisions The Hidden Biases in AI Recruitment: Addressing Systemic Discrimination in Hiring AlgorithmsIntroductionAI-powered hiring tools promise to transform recruitment with efficient candidate screening, standardized interview processes, and data-driven sel
Corporate AI Adoption Plateaus, Ramp Data Reveals Corporate AI Adoption Reaches PlateauWhile businesses initially rushed to implement artificial intelligence solutions, enthusiasm appears to be stabilizing as organizations confront the technology's current limitations.The Adoption SlowdownRamp's AI
Pokemon FireRed Kaizo IronMon Challenge: Essential Rules & Winning Strategies The Pokemon FireRed Kaizo IronMon challenge stands as one of gaming's ultimate tests of skill—a brutal gauntlet that breaks conventional Pokemon strategies and forces players to rethink every decision. This punishing variant combines ruthless randomi
AI-Driven Task Management Tools Maximize Productivity and Efficiency The Future of Productivity: AI-Powered Task ManagementIn our constantly accelerating digital landscape, effective task management has become essential for professional success. Artificial intelligence is revolutionizing how we organize workflows, bri
Model comparison
Start the comparison
Back to Top
OR