Model Introduction
The Llama 3.2 3B models support context length of 128K tokens and are state-of-the-art in their class for on-device use cases like summarization, instruction following, and rewriting tasks running locally at the edge.
Language comprehension ability
Often makes semantic misjudgments, leading to obvious logical disconnects in responses.
3.1
Knowledge coverage scope
Has significant knowledge blind spots, often showing factual errors and repeating outdated information.
4.2
Reasoning ability
Unable to maintain coherent reasoning chains, often causing inverted causality or miscalculations.
3.0