Alibaba/Mixture of Experts

AlibabaQwen 3.6 35B A3B

Qwen 3.6 35B A3B — hybrid linear/full attention MoE (DeltaNet + full attention), multimodal (text+image+video), 256 experts (8+1 active). Prioritizes agentic coding and thinking preservation over Qwen 3.5.

chatcodingreasoningmultilingualvisionmathtool_use
35B
Parameters (3B active)
256K
Context length
14
Benchmarks
14
Quantizations
100K
HF downloads
Architecture
MoE
Released
2026-04-15
Layers
40
KV Heads
2
Head Dim
256
Family
qwen

Quantization Options

QuantBitsVRAMQuality
IQ3_XXS3.2514.7 GBlow
IQ3_XS3.515.8 GBlow
Q3_K_S3.6416.4 GBlow
IQ3_M3.7216.8 GBlow
Q3_K_M418.0 GBlow
Q3_K_L4.2519.1 GBlow
IQ4_XS4.3719.6 GBmoderate
Q4_K_S4.520.2 GBmoderate
Q4_K_M4.8921.9 GBgood
Q5_K_S5.5724.9 GBgood
Q5_K_M5.725.4 GBgood
Q6_K6.5629.2 GBexcellent
Q8_08.537.7 GBlossless
FP161670.5 GBlossless

Select your GPU above to see speed estimates and compatibility for each quantization.

Community Ratings

Loading ratings...

Benchmarks (14)

MMBench92.8
AIME92.7
GPQA Diamond86.0
MMLU-PRO85.2
MMMU81.7
LiveCodeBench80.4
IFEval78.9
SWE-bench73.4
MATH59.7
BBH58.3
BigCodeBench32.3
HLE21.4
MUSR19.1
GPQA15.2

Run this model

Easiest way to get started·Beginners
DOCS ↗
curl -fsSL https://ollama.com/install.sh | sh
$ollama run qwen3.6:35b-a3b:q4_k_m

Downloads and runs automatically. Add --verbose for speed stats.

▸ SETUP GUIDE
>_

Auto-setup with fitmyllm CLI

Detects your GPU, recommends the best model, downloads it, and starts chatting — zero config. Benchmarks your speed and contributes anonymous data to improve predictions.

pip install fitmyllmthen run fitmyllmLearn more
Auto-detect GPULive tok/s in chatSpeed benchmarks9 inference engines

GPUs that can run this model

At Q4_K_M quantization. Sorted by minimum VRAM.

Find the best GPU for Qwen 3.6 35B A3B

Build Hardware for Qwen 3.6 35B A3B
▸ SPEC SHEET

Qwen 3.6 35B A3B35B MoE.

▸ SPECIFICATIONS
PARAMETERS
35B (3B active)
ARCHITECTURE
Mixture of Experts
CONTEXT LENGTH
256K tokens
CAPABILITIES
chat, coding, reasoning, multilingual, vision, math, tool_use
RELEASE DATE
2026-04-15
PROVIDER
Alibaba
FAMILY
qwen
▸ VRAM REQUIREMENTS
QUANTBPWVRAMQUALITY
IQ3_XXS3.2514.7 GB78%
IQ3_XS3.515.8 GB82%
Q3_K_S3.6416.4 GB83%
IQ3_M3.7216.8 GB84%
Q3_K_M418.0 GB88%
Q3_K_L4.2519.1 GB89%
IQ4_XS4.3719.6 GB91%
Q4_K_S4.520.2 GB92%
Q4_K_M4.8921.9 GB94%
Q5_K_S5.5724.9 GB96%
Q5_K_M5.725.4 GB96%
Q6_K6.5629.2 GB97%
Q8_08.537.7 GB100%
FP161670.5 GB100%
§ 01BENCHMARK SCORES
MMLU-PRO85.2
MATH59.7
IFEval78.9
BBH58.3
GPQA15.2
MUSR19.1
BigCodeBench32.3
LiveCodeBench80.4
SWE-bench73.4
AIME92.7
GPQA Diamond86.0
HLE21.4
MMMU81.7
MMBench92.8
§ 02RUN COMMAND

Run Qwen 3.6 35B A3B locally with Ollama — needs 21.9 GB VRAM at Q4_K_M:

$ollama run qwen3.6:35b-a3b