OpenAI/Dense

OpenAIGPT-2 Large 774M

Larger GPT-2. Can generate coherent text but outdated by modern standards.

chatThinking
0.774B
Parameters
1K
Context length
6
Benchmarks
6
Quantizations
500K
HF downloads
Architecture
Dense
Released
2019-02-14
Layers
36
KV Heads
20
Head Dim
64
Family
gpt2

Quantization Options

QuantBitsVRAMQuality
Q4_K_M4.891.0 GBgood
Q5_K_S5.571.0 GBgood
Q5_K_M5.71.0 GBgood
Q6_K6.561.1 GBexcellent
Q8_08.51.3 GBlossless
FP16162.0 GBlossless

Select your GPU above to see speed estimates and compatibility for each quantization.

READY TO RUN THIS?RENT BY THE HOUR

RENT A GPU AND RUN GPT-2 LARGE 774M NOW

Spin up an A100 / H100 / 4090 in ~60s. Pay by the second. Cancel anytime.

Community Ratings

Loading ratings...

Benchmarks (6)

IFEval20.5
MUSR5.7
BBH3.3
MMLU-PRO1.6
MATH1.2
GPQA1.2

Run this model

Easiest way to get started·Beginners
DOCS ↗
curl -fsSL https://ollama.com/install.sh | sh
$ollama run gpt2:0.774b-q4_K_M

Tag may need adjustment — check ollama.com/library/gpt2 for available tags.

▸ SETUP GUIDE
>_

Auto-setup with fitmyllm CLI

Detects your GPU, recommends the best model, downloads it, and starts chatting — zero config. Benchmarks your speed and contributes anonymous data to improve predictions.

pip install fitmyllmthen run fitmyllmLearn more
Auto-detect GPULive tok/s in chatSpeed benchmarks9 inference engines

GPUs that can run this model

At Q4_K_M quantization. Sorted by minimum VRAM.

Find the best GPU for GPT-2 Large 774M

Build Hardware for GPT-2 Large 774M

Larger GPT-2. Can generate coherent text but outdated by modern standards.

▸ SPEC SHEET

GPT-2 Large 774M0.774B Dense.

▸ SPECIFICATIONS
PARAMETERS
0.774B
ARCHITECTURE
Dense Transformer
CONTEXT LENGTH
1K tokens
CAPABILITIES
chat
RELEASE DATE
2019-02-14
PROVIDER
OpenAI
FAMILY
gpt2
▸ VRAM REQUIREMENTS
QUANTBPWVRAMQUALITY
Q4_K_M4.891.0 GB94%
Q5_K_S5.571.0 GB96%
Q5_K_M5.71.0 GB96%
Q6_K6.561.1 GB97%
Q8_08.51.3 GB100%
FP16162.0 GB100%
§ 01BENCHMARK SCORES
MMLU-PRO1.6
MATH1.2
IFEval20.5
BBH3.3
GPQA1.2
MUSR5.7