EXAONE Deep 7.8B — 7.8B Parameter Dense LLM
Model Specifications
- Parameters
- 7.8B
- Architecture
- Dense Transformer
- Context Length
- 32K tokens
- Capabilities
- reasoning, math, coding
- Release Date
- 2025-03-16
- Family
- exaone
VRAM Requirements
| Quantization | BPW | VRAM | Quality |
|---|---|---|---|
| Q4_K_M | 4.89 | 5.3 GB | 94% |
| Q5_K_S | 5.57 | 5.9 GB | 96% |
| Q5_K_M | 5.7 | 6.0 GB | 96% |
| Q6_K | 6.56 | 6.9 GB | 97% |
| Q8_0 | 8.5 | 8.8 GB | 100% |
| FP16 | 16 | 16.1 GB | 100% |
Benchmark Scores
MMLU-PRO34.8
MATH47.5
IFEval81.4
BBH25.7
GPQA1.0
MUSR4.9
MATH-50094.8
How to Run EXAONE Deep 7.8B
Run EXAONE Deep 7.8B locally with Ollama (needs 5.3 GB VRAM at Q4_K_M):
ollama run exaone-deep:7.8bCompatible GPUs (30)
GPUs that can run EXAONE Deep 7.8B at Q4_K_M quantization:
NVIDIA RTX 3050 6GB(6GB, 168 GB/s)Intel Arc A380(6GB, 186 GB/s)NVIDIA RTX 2060 6GB(6GB, 336 GB/s)NVIDIA GTX 1660 SUPER(6GB, 336 GB/s)NVIDIA GTX 1660 Ti(6GB, 288 GB/s)NVIDIA GTX 1060 6GB(6GB, 192 GB/s)NVIDIA Tesla C2070(6GB, 143 GB/s)NVIDIA Tesla C2075(6GB, 150 GB/s)NVIDIA Tesla C2090(6GB, 177 GB/s)NVIDIA Tesla M2070(6GB, 150 GB/s)NVIDIA Tesla M2070-Q(6GB, 150 GB/s)NVIDIA Tesla M2075(6GB, 150 GB/s)NVIDIA Tesla M2090(6GB, 177 GB/s)NVIDIA Tesla X2070(6GB, 177 GB/s)NVIDIA Tesla X2090(6GB, 177 GB/s)