On March 2, 2026, Alibaba open-sourced the Qwen3.5 small model series. The 9B version achieves an impressive 81.7 on GPQA Diamond, outperforming OpenAI's GPT-OSS-120B (71.5). Despite a 13.5x parameter gap, the small model wins.
Released under Apache 2.0 license, both code and weights are available for commercial use. Deploy with a single Ollama command on standard laptops.
Figure 1: Qwen3.5 Small Model Performance Comparison (Source: GitHub README)
1. Qwen3.5 Small Model Series
On March 2, 2026, Alibaba Qianwen team open-sourced four Qwen3.5 small models: Qwen3.5-0.8B, 2B, 4B, and 9B.
Figure 2: Qwen3.5 Medium Model Performance (Source: GitHub README)
These are not "shrunk versions". This series uses native multimodal training with the latest model architecture.
Model Positioning
| Model | Positioning | Features | Use Cases |
|---|---|---|---|
| 0.8B/2B | Edge Device First | Ultra-small, ultra-fast inference | Mobile devices, IoT, real-time interaction |
| 4B | Lightweight Agent | Multimodal base | Agent core |
| 9B | Compact Size, Beyond-Level Performance | Competes with 120B | Server-side, memory-constrained |
The 0.8B and 2B models are suitable for mobile devices and IoT edge deployment. The 4B model fits lightweight agents. The 9B model is ideal for server-side deployment with exceptional cost-effectiveness.
2. 9B vs 120B: Benchmark Results
GPQA Diamond benchmark results:
| Model | GPQA Diamond | Parameters | Type |
|---|---|---|---|
| Qwen3.5-9B | 81.7 | 9B | End-to-End |
| GPT-OSS-120B | 71.5 | 120B | End-to-End |
Qwen3.5-9B outperforms GPT-OSS-120B by 10.2 points.
VentureBeat's headline was direct: "Alibaba's small, open source Qwen3.5-9B beats OpenAI's gpt-oss-120B and can run on standard laptops".
What does "can run on standard laptops" mean? The 9B model uses approximately 4-5GB VRAM. RTX 3090, A10, or even high-end laptop GPUs can run it. No need for A100 or H100 datacenter-grade GPUs.
Previously, running a 120B model required at least 8 A100s. Now, the 9B model runs on a single card. The cost difference is orders of magnitude.
3. Technical Highlights: Why Can Small Models Win?
Qwen3.5 is not "distillation" or "pruning". There are several technical breakthroughs:
Unified Vision-Language Foundation
Early fusion training with trillions of multimodal tokens. Qwen3.5 surpasses Qwen3-VL models in reasoning, encoding, agent capabilities, and multimodal understanding.
Figure 3: Qwen3.5 Flagship Model Performance Comparison (Source: GitHub README)
Efficient Hybrid Architecture
Gated Delta Networks combined with sparse MoE (Mixture-of-Experts). High-throughput inference with low latency.
Qwen3.5-397B-A17B has 397B total parameters but only activates 17B per forward pass. Qwen3.5-9B doesn't disclose its MoE configuration but inherits the same architectural philosophy.
Scalable RL Generalization
Scaling reinforcement learning in millions of agent environments. Not optimization for specific benchmarks, but real-world adaptability.
Global Language Coverage
Expanded from 119 languages to 201. Vocabulary grew from 150k to 250k, improving encoding/decoding efficiency by 10-60%.
4. Deployment: One Command
How simple is deploying Qwen3.5-9B? One Ollama command:
ollama run qwen3.5:9b
That's it.
Using transformers:
from transformers import Qwen3VLForConditionalGeneration, AutoProcessor
model = Qwen3VLForConditionalGeneration.from_pretrained(
"Qwen/Qwen3.5-9B",
torch_dtype=torch.bfloat16,
device_map="auto",
)
processor = AutoProcessor.from_pretrained("Qwen/Qwen3.5-9B")
VRAM Usage
- bfloat16 precision: ~4-5GB
- int8 quantization: ~2-3GB
- int4 quantization: ~1-2GB
Inference Speed (Single RTX 3090)
- Generation speed: ~30-50 tokens/second
- First token latency: <100ms
Comparison with 120B model:
- VRAM usage: ~240GB (bfloat16)
- Requires: 8× A100 (80GB each)
- Inference speed: ~5-10 tokens/second
5. Selection Guide: How to Choose 0.8B/2B/4B/9B?
| Requirement | Recommended Model | Reason |
|---|---|---|
| Mobile Deployment | 0.8B/2B | Ultra-small, ultra-fast |
| IOT Edge Devices | 0.8B/2B | Low resource consumption |
| Lightweight Agent | 4B | Balanced performance/resources |
| Server-General Purpose | 9B | Best cost-performance |
| VRAM <4GB | 0.8B/2B | Minimum resource requirement |
| VRAM 4-8GB | 4B/9B | Medium resource requirement |
| Maximum Performance | 9B | Near 120B performance |
Recommendations
- Ample VRAM (≥8GB): Go straight for 9B
- Mobile development: Choose 2B
- Agent development: 4B is the sweet spot
6. Conclusion: The Era of Small Models
The open-source release of Qwen3.5-9B marks a new trend: small models are no longer a "compromise" but a "choice".
Previously, performance = parameters. The fact that 9B surpasses 120B tells us: architecture optimization > parameter stacking.
This is great news for developers. Previously limited to cloud API calls, now local deployment is possible. Previously concerned about data privacy, now fully offline operation is feasible. Previously too costly, now a single card handles it.