Rednote/Dense

Rdots.llm1.inst 142.8B

chat
142.8B
Parameters
32K
Context length
0
Benchmarks
17
Quantizations
8K
HF downloads
Architecture
Dense
Released
2025-05-14
Layers
62
KV Heads
32
Head Dim
128
Family
other

Quantization Options

QuantBitsVRAMQuality
IQ2_XXS2.3843.0 GBlow
IQ2_M2.9352.8 GBlow
Q2_K3.1656.9 GBlow
IQ3_XXS3.2558.5 GBlow
IQ3_XS3.563.0 GBlow
Q3_K_S3.6465.5 GBlow
IQ3_M3.7667.6 GBlow
Q3_K_M471.9 GBlow
Q3_K_L4.377.2 GBmoderate
IQ4_XS4.4680.1 GBmoderate
Q4_K_S4.6783.8 GBmoderate
Q4_K_M4.8987.8 GBgood
Q5_K_S5.5799.9 GBgood
Q5_K_M5.7102.2 GBgood
Q6_K6.56117.6 GBexcellent
Q8_08.5152.2 GBlossless
FP1616286.1 GBlossless

Select your GPU above to see speed estimates and compatibility for each quantization.

Run this model

Easiest way to get starteddocs →
curl -fsSL https://ollama.com/install.sh | sh
$ollama run other:142b-q4_k_m

Downloads and runs automatically. Add --verbose for speed stats.

Setup guide

GPUs that can run this model

At Q4_K_M quantization. Sorted by minimum VRAM.

NVIDIA B200
90 GB VRAM • 4100 GB/s
NVIDIA
$30000
NVIDIA H100 NVL 94 GB
94 GB VRAM • 3940 GB/s
NVIDIA
$30000
NVIDIA H100 SXM5 94 GB
94 GB VRAM • 3360 GB/s
NVIDIA
$25000
RTX Pro 6000
96 GB VRAM • 1792 GB/s
NVIDIA
$8565
NVIDIA H100 PCIe 96 GB
96 GB VRAM • 3360 GB/s
NVIDIA
$25000
NVIDIA H100 SXM5 96 GB
96 GB VRAM • 3360 GB/s
NVIDIA
$25000
Intel Data Center GPU Max 1350
96 GB VRAM • 2460 GB/s
INTEL
NVIDIA RTX PRO 6000 Blackwell Server
96 GB VRAM • 1790 GB/s
NVIDIA
$9999
NVIDIA RTX PRO 6000 Blackwell
96 GB VRAM • 1790 GB/s
NVIDIA
$9999
AMD Instinct MI300A
120 GB VRAM • 5300 GB/s
AMD
$12000
Apple M4 Max (128GB)
128 GB VRAM • 546 GB/s
APPLE
$3999
AMD Instinct MI250X
128 GB VRAM • 3277 GB/s
AMD
$10000
Apple M1 Ultra (128GB)
128 GB VRAM • 800 GB/s
APPLE
$4999
Apple M2 Ultra (128GB)
128 GB VRAM • 800 GB/s
APPLE
$3999
AMD Radeon Instinct MI250
128 GB VRAM • 3280 GB/s
AMD
$12000
AMD Radeon Instinct MI250X
128 GB VRAM • 3280 GB/s
AMD
$15000
AMD Radeon Instinct MI300
128 GB VRAM • 6550 GB/s
AMD
$12000
Intel Data Center GPU Max 1550
128 GB VRAM • 3280 GB/s
INTEL
Intel Data Center GPU Max Subsystem
128 GB VRAM • 3210 GB/s
INTEL
NVIDIA GB10
128 GB VRAM • 273 GB/s
NVIDIA
NVIDIA Jetson T5000
128 GB VRAM • 273 GB/s
NVIDIA
Apple M5 Max (128GB)
128 GB VRAM • 614 GB/s
APPLE
NVIDIA H200 SXM 141GB
140 GB VRAM • 4800 GB/s
NVIDIA
$30000
NVIDIA H200 NVL
141 GB VRAM • 4890 GB/s
NVIDIA
$35000
NVIDIA H200 SXM 141 GB
141 GB VRAM • 4890 GB/s
NVIDIA
$30000
NVIDIA B300
144 GB VRAM • 4100 GB/s
NVIDIA
$35000
AMD Instinct MI300X
192 GB VRAM • 5300 GB/s
AMD
$15000
Apple M2 Ultra (192GB)
192 GB VRAM • 800 GB/s
APPLE
$5499
Apple M3 Ultra (192GB)
192 GB VRAM • 800 GB/s
APPLE
$6999
Apple M4 Ultra (192GB)
192 GB VRAM • 1092 GB/s
APPLE
$7499

Find the best GPU for dots.llm1.inst 142.8B

Build Hardware for dots.llm1.inst 142.8B

dots.llm1.inst 142.8B142.8B Parameter Dense LLM

Model Specifications

Parameters
142.8B
Architecture
Dense Transformer
Context Length
32K tokens
Capabilities
chat
Release Date
2025-05-14
Provider
Rednote
Family
other

VRAM Requirements

QuantizationBPWVRAMQuality
IQ2_XXS2.3843.0 GB65%
IQ2_M2.9352.8 GB75%
Q2_K3.1656.9 GB78%
IQ3_XXS3.2558.5 GB82%
IQ3_XS3.563.0 GB84%
Q3_K_S3.6465.5 GB85%
IQ3_M3.7667.6 GB86%
Q3_K_M471.9 GB88%
Q3_K_L4.377.2 GB90%
IQ4_XS4.4680.1 GB92%
Q4_K_S4.6783.8 GB93%
Q4_K_M4.8987.8 GB94%
Q5_K_S5.5799.9 GB96%
Q5_K_M5.7102.2 GB96%
Q6_K6.56117.6 GB97%
Q8_08.5152.2 GB100%
FP1616286.1 GB100%

How to Run dots.llm1.inst 142.8B

Run dots.llm1.inst 142.8B locally with Ollama (needs 87.8 GB VRAM at Q4_K_M):

ollama run other:142b

Compatible GPUs (30)

GPUs that can run dots.llm1.inst 142.8B at Q4_K_M quantization: