▸ DEVICE UNDER TEST
ARM Mali-G72 MP18 — 6 GB VRAM.
▸ ARM MALI-G72 MP18 SPEC
- BRAND
- Apple
- VRAM
- 6 GB Shared
- BANDWIDTH
- 29.8 GB/s
- FP16 COMPUTE
- 0.5 TFLOPS
- TDP
- 5 W
- ARCHITECTURE
- Bifrost Gen2
▸ AI CAPABILITY
142/ 331 models @ Q4
With 6 GB VRAM and 29.8 GB/s bandwidth, this GPU handles models up to 8B parameters.
Speed ≈ bandwidth / model_size × efficiency. A 7B model at Q4 runs at ~3 tok/s.
§ 01TOP MODELS FOR ARM MALI-G72 MP18
142 FIT · SHOWING 20| MODEL | SIZE | VRAM Q4 | TOK/S | AVG |
|---|---|---|---|---|
| Aya Expanse 8B | 8B | 5.4 GB | 3 | 27.8 |
| Cogito 8B | 8B | 5.4 GB | 3 | 17.2 |
| DeepSeek R1 Distill Llama 8B | 8B | 5.4 GB | 3 | 36.2 |
| Gemma 3n E4B | 8B | 5.4 GB | 3 | 28.8 |
| Granite 3.3 8B | 8B | 5.4 GB | 3 | 24.6 |
| Llama-3.1-8B | 8B | 5.4 GB | 3 | 23.5 |
| Dolphin Llama 3 8B | 8B | 5.4 GB | 3 | 23.8 |
| Llama 3 8B | 8B | 5.4 GB | 3 | 37.3 |
| Tulu 3 8B | 8B | 5.4 GB | 3 | 31.5 |
| Ministral-8B | 8B | 5.4 GB | 3 | 19.3 |
| Nemotron-H 8B | 8B | 5.4 GB | 3 | 78.4 |
| Granite 8B | 8B | 5.4 GB | 3 | 26.1 |
| InternVL2 8B | 8B | 5.4 GB | 3 | 44.6 |
| MiniCPM-V 2.6 8B | 8B | 5.4 GB | 3 | 40.8 |
| RNJ-1 8B | 8B | 5.4 GB | 3 | 53.5 |
| Gemma 4 E4B | 8B | 5.4 GB | 3 | 32.1 |
| EXAONE Deep 7.8B | 7.8B | 5.3 GB | 3 | 41.4 |
| InternLM2.5 7B | 7.74B | 5.2 GB | 3 | 44.8 |
| Qwen2.5-7B | 7.6B | 5.1 GB | 3 | 35.2 |
| Qwen2.5-Coder-7B | 7.6B | 5.1 GB | 3 | 31.0 |