Qwen/Qwen3.5-2B
About
Released: 2/25/2026Qwen/Qwen3.5-2B is a multimodal LLM with native vision capabilities. It excels in general-purpose dialogue, instruction following, coding, mathematics, multilingual tasks, and vision-language understanding, with strong performance on reasoning, agents, and visual understanding benchmarks. The 2B parameter scale makes it suitable for prototyping, task-specific fine-tuning, and resource-constrained deployment.
Key features include a unified vision-language foundation with early fusion training, support for 201 languages and dialects, long-context processing up to 262,144 tokens, image and video understanding, tool calling for agentic use cases, optional thinking mode for complex reasoning, and a hybrid architecture using Gated Delta Networks with sparse Mixture-of-Experts.
| Metric | Value |
|---|---|
| Parameter Count | 2 billion |
| Mixture of Experts | No |
| Context Length | 262,144 tokens |
| Multilingual | Yes |
| Quantized* | No |
*Quantization is specific to the inference provider and the model may be offered with different quantization levels by other providers.