qwen

Qwen/Qwen3.5-2B

Fine-tunable$0.0014/sec
multi-to-text
About
Released: 2/25/2026

Qwen/Qwen3.5-2B is a multimodal LLM with native vision capabilities. It excels in general-purpose dialogue, instruction following, coding, mathematics, multilingual tasks, and vision-language understanding, with strong performance on reasoning, agents, and visual understanding benchmarks. The 2B parameter scale makes it suitable for prototyping, task-specific fine-tuning, and resource-constrained deployment.

Key features include a unified vision-language foundation with early fusion training, support for 201 languages and dialects, long-context processing up to 262,144 tokens, image and video understanding, tool calling for agentic use cases, optional thinking mode for complex reasoning, and a hybrid architecture using Gated Delta Networks with sparse Mixture-of-Experts.

MetricValue
Parameter Count2 billion
Mixture of ExpertsNo
Context Length262,144 tokens
MultilingualYes
Quantized*No

*Quantization is specific to the inference provider and the model may be offered with different quantization levels by other providers.