Model Introduction
Qwen 2.5 Max is a large-scale MoE (Mixture-of-Experts) model trained with over 20 trillion tokens of pre-training data and a meticulously designed post-training scheme.
Language comprehension ability
Capable of understanding complex contexts and generating logically coherent sentences, though occasionally off in tone control.
7.5
Knowledge coverage scope
Possesses core knowledge of mainstream disciplines, but has limited coverage of cutting-edge interdisciplinary fields.
8.8
Reasoning ability
Unable to maintain coherent reasoning chains, often causing inverted causality or miscalculations.
6.8