Can NVIDIA GeForce RTX 4060 Ti 8GB run Qwen 2.5 72B?
72B parameter LLM model on 8GB GDDR6
No — not enough VRAM
SpeedWill not load
QualityN/A
VRAM Requirements
Qwen 2.5 72B is a 72B parameter model. At full precision (FP16), it requires 144GB of VRAM. Your NVIDIA GeForce RTX 4060 Ti 8GB only has 8GB — not enough even at maximum compression.
FP16 (Full Precision)144GB (need 136GB more)
Maximum quality, no quantization
Q8 (8-bit)72GB (need 64GB more)
Near-lossless, ~50% size reduction
Q4 (4-bit)42GB (need 34GB more)
Good quality, ~75% size reduction
Your GPU VRAM: 8GB GDDR6 at 288 GB/s bandwidth
Recommended system RAM: 32GB DDR5 (2x GPU VRAM minimum for model overflow)
Recommended system RAM: 32GB DDR5 (2x GPU VRAM minimum for model overflow)
NVIDIA GeForce RTX 4060 Ti 8GB Specs
VRAM8GB GDDR6
Memory Bandwidth288 GB/s
TDP160W
CUDA Cores4,352
Street Price~$370
AI Rating3/10
Other GPUs That Run Qwen 2.5 72B
Other LLM Models on NVIDIA GeForce RTX 4060 Ti 8GB
About Qwen 2.5 72B
Top open LLM for reasoning. Similar requirements to Llama 70B.
Category: LLM · Parameters: 72B · CUDA required: No (runs via llama.cpp/GGUF)