← Back to directory
Alibaba · Released 2026-04
Qwen3.6 7B
7B variant of Qwen3.6 with vision support. Punches above its weight class, with the broadest language coverage of any 7B model.
QwenCommercial use OK👁 Visionsmallvision
Params (max)
7B
Variants
7B
Context window
128K tokens
MMLU
76.4
HumanEval
75
GSM8K
84.6
Min VRAM (fp16, smallest variant)
16GB
Smallest Q4 GGUF
~4.5GB
Languages supported
119
Pros
- ✓Best multilingual small model
- ✓Vision included
- ✓Permissive license
Cons
- ×Smaller fine-tune community
- ×Vision tuned heavier on Chinese
Highlights
- ●Vision in 7B
- ●119 languages
- ●Strong code at small size
Where to download
Hugging Face: Qwen/Qwen3.6-7B-Instruct
Or via Ollama (
ollama pull qwen-3.6-7b) or LM Studio's in-app browser.Homepage: https://qwenlm.ai
Related reading
Running an LLM on Your Laptop in 2026: M-Series, Quantization, and What Actually Works
Step-by-step: pick a quantization, install Ollama or LM Studio, run a 7B-14B model on a MacBook or 16GB GPU, and not lose your sanity.
Small LLMs on Edge Devices: What Runs on Phones, Pis, and Browsers in 2026
Gemma 2B runs on a Pi 5. Phi-4 runs in a browser via WebGPU. Phones run Llama 3B. A practical guide to LLMs on tiny hardware.