option
Home
List of Al models
DeepSeek-V2-Chat-0628

DeepSeek-V2-Chat-0628

Add comparison
Add comparison
Model parameter quantity
236B
Model parameter quantity
Affiliated organization
DeepSeek
Affiliated organization
Open Source
License Type
Release time
May 6, 2024
Release time

Model Introduction
DeepSeek-V2 is a strong Mixture-of-Experts (MoE) language model characterized by economical training and efficient inference. It comprises 236B total parameters, of which 21B are activated for each token. Compared with DeepSeek 67B, DeepSeek-V2 achieves stronger performance, and meanwhile saves 42.5% of training costs, reduces the KV cache by 93.3%, and boosts the maximum generation throughput to 5.76 times.
Swipe left and right to view more
Language comprehension ability Language comprehension ability
Language comprehension ability
Often makes semantic misjudgments, leading to obvious logical disconnects in responses.
4.6
Knowledge coverage scope Knowledge coverage scope
Knowledge coverage scope
Possesses core knowledge of mainstream disciplines, but has limited coverage of cutting-edge interdisciplinary fields.
7.8
Reasoning ability Reasoning ability
Reasoning ability
Unable to maintain coherent reasoning chains, often causing inverted causality or miscalculations.
4.7
Related model
DeepSeek-V3-0324 DeepSeek-V3 outperforms other open-source models such as Qwen2.5-72B and Llama-3.1-405B in multiple evaluations and matches the performance of top-tier closed-source models like GPT-4 and Claude-3.5-Sonnet.
DeepSeek-R1-0528 The latest version of Deepseek R1.
DeepSeek-V2.5 DeepSeek-V2.5 is an upgraded version that combines DeepSeek-V2-Chat and DeepSeek-Coder-V2-Instruct. The new model integrates the general and coding abilities of the two previous versions.
DeepSeek-V3-0324 DeepSeek-V3 outperforms other open-source models such as Qwen2.5-72B and Llama-3.1-405B in multiple evaluations and matches the performance of top-tier closed-source models like GPT-4 and Claude-3.5-Sonnet.
DeepSeek-V2-Lite-Chat DeepSeek-V2, a strong Mixture-of-Experts (MoE) language model presented by DeepSeek, DeepSeek-V2-Lite is a lite version of it.
Relevant documents
Google Search Expands Smarter AI Mode Worldwide Google is bringing its AI-powered search experience to 180 additional countries, significantly expanding beyond its initial US, UK, and India rollout. While currently English-only, this global expansion enables more users worldwide to experience conv
Step-by-Step Guide to Creating Amazon Coloring Books Using Leonardo AI Dreaming of breaking into Amazon's thriving book market? Coloring books offer a fantastic passive income opportunity, but finding distinctive artwork can be difficult. This comprehensive tutorial reveals how Leonardo AI can help you craft captivating
Amazon introduces AI-powered Lens Live for real-world shopping experiences Amazon continues advancing AI-powered shopping innovations with Tuesday's introduction of Lens Live, an enhanced version of its visual search technology that gives customers real-time product discovery capabilities. This upgrade builds upon Amazon Le
"AI Mode Introduces Innovative Ways to Engage with Information" The Future of AI-Powered Search Is Here: AI Mode Opens to EveryoneMillions of users are revolutionizing their search experience with AI Mode in Google Labs - tackling complex inquiries, refining results through conversational follow-ups, and uncoveri
Easily Create Stunning Comics with AI-Powered Comic Factory Comic Factory AI represents a quantum leap in digital storytelling, transforming how visual narratives are crafted. This groundbreaking platform leverages artificial intelligence to instantly convert text descriptions into professional-grade comic ar
Model comparison
Start the comparison
Back to Top
OR