LLaVA Team/Dense
LLaVA-1.6 Yi 34B
visionchat
34B
Parameters
4K
Context length
8
Benchmarks
4
Quantizations
120K
HF downloads
Architecture
Dense
Released
2024-01-30
Layers
60
KV Heads
8
Head Dim
128
Family
other
Model Card
View on HuggingFaceLLaVA Model Card
Model details
Model type: LLaVA is an open-source chatbot trained by fine-tuning LLM on multimodal instruction-following data. It is an auto-regressive language model, based on the transformer architecture. Base LLM: NousResearch/Nous-Hermes-2-Yi-34B
Model date: LLaVA-v1.6-34B was trained in December 2023.
Paper or resources for more information: https://llava-vl.github.io/
Quantizations & VRAM
Q4_K_M4.5 bpw
20.5 GB
VRAM required
94%
Quality
Q6_K6.5 bpw
27.3 GB
VRAM required
97%
Quality
Q8_08 bpw
35.2 GB
VRAM required
100%
Quality
FP1616 bpw
68.4 GB
VRAM required
100%
Quality
Benchmarks (8)
MMBench79.3
IFEval68.0
MMLU-PRO58.0
BBH58.0
MMMU46.7
GPQA35.0
MUSR18.0
MATH16.5
Run with Ollama
$
ollama run llava:34bGPUs that can run this model
At Q4_K_M quantization. Sorted by minimum VRAM.
NVIDIA RTX 4090
24 GB VRAM • 1008 GB/s
NVIDIA
$1599
NVIDIA RTX 3090 Ti
24 GB VRAM • 1008 GB/s
NVIDIA
$999
NVIDIA RTX 3090
24 GB VRAM • 936 GB/s
NVIDIA
$850
AMD RX 7900 XTX
24 GB VRAM • 960 GB/s
AMD
$999
Apple M4 Pro (24GB)
24 GB VRAM • 273 GB/s
APPLE
$1399
NVIDIA L4 24GB
24 GB VRAM • 300 GB/s
NVIDIA
$2500
NVIDIA A10 24GB
24 GB VRAM • 600 GB/s
NVIDIA
$3500
Apple M2 (24GB)
24 GB VRAM • 100 GB/s
APPLE
$999
Apple M3 (24GB)
24 GB VRAM • 100 GB/s
APPLE
$999
Apple M4 (24GB)
24 GB VRAM • 120 GB/s
APPLE
$699
NVIDIA Tesla M40 24 GB
24 GB VRAM • 288 GB/s
NVIDIA
NVIDIA Tesla P10
24 GB VRAM • 694 GB/s
NVIDIA
NVIDIA Tesla P40
24 GB VRAM • 347 GB/s
NVIDIA
NVIDIA Quadro RTX 6000
24 GB VRAM • 672 GB/s
NVIDIA
NVIDIA Quadro RTX 6000 Passive
24 GB VRAM • 624 GB/s
NVIDIA
NVIDIA GeForce RTX 3090
24 GB VRAM • 936 GB/s
NVIDIA
$1499
NVIDIA A10 PCIe
24 GB VRAM • 600 GB/s
NVIDIA
NVIDIA A10G
24 GB VRAM • 600 GB/s
NVIDIA
NVIDIA RTX A5000
24 GB VRAM • 768 GB/s
NVIDIA
NVIDIA GeForce RTX 3090 Ti
24 GB VRAM • 1010 GB/s
NVIDIA
$1999
NVIDIA GeForce RTX 4090
24 GB VRAM • 1010 GB/s
NVIDIA
$1599
NVIDIA L40 CNX
24 GB VRAM • 864 GB/s
NVIDIA
NVIDIA L40G
24 GB VRAM • 864 GB/s
NVIDIA
AMD Radeon RX 7900 XTX
24 GB VRAM • 960 GB/s
AMD
$999
NVIDIA GeForce RTX 4090 D
24 GB VRAM • 1010 GB/s
NVIDIA
$1599
NVIDIA GeForce RTX 5090 D V2
24 GB VRAM • 1340 GB/s
NVIDIA
$1999
NVIDIA TITAN RTX
24 GB VRAM • 672 GB/s
NVIDIA
NVIDIA A30 PCIe
24 GB VRAM • 933 GB/s
NVIDIA
NVIDIA A30X
24 GB VRAM • 1220 GB/s
NVIDIA
NVIDIA PG506-207
24 GB VRAM • 933 GB/s
NVIDIA
Find the best GPU for LLaVA-1.6 Yi 34B
Build Hardware for LLaVA-1.6 Yi 34B