← Back to GPUs

NVIDIA · RTX 40
NVIDIA GeForce RTX 4060
$280$299 MSRP
The RTX 4060 is a budget-friendly 1080p gaming card with just 8GB of GDDR6 memory. For gaming at 1080p, it performs well with DLSS 3 support. However, the 8GB VRAM severely limits AI capability — you are restricted to 7B models with heavy quantization and basic Stable Diffusion 1.5. If AI is any part of your plans, spend the extra on the 4060 Ti 16GB instead.
Best ForBudget 1080p gaming only
VerdictFine for gaming, but 8GB is not enough for meaningful AI work.
AI
3/10
Gaming
6/10
Specifications
VRAM8GB GDDR6
Memory Bandwidth272 GB/s
CUDA Cores3,072
Boost Clock2460 MHz
TDP115W
Power Connector1x 8-pin
Length240mm
Form FactorDual Slot
Release Year2023
AI Capabilities
Entry Level8GB VRAM
Limited to small models with heavy quantization. Fine for experimenting.
Can run (Q4 quantized)
Llama 3.1 8BMistral 7BStable Diffusion XLStable Diffusion 3.5 LargeCogVideoX-5BLTX VideoStable Video DiffusionAlphaFold 2ESM-2 3BscGPTRFdiffusionFine-tune Llama 8BTrain SDXL LoRA
Tight fit (may need CPU offload)
Qwen 2.5 14B (9GB Q4)FLUX.1 Dev (10GB Q4)Mochi 1 (10GB Q4)Wan Video 14B (11GB Q4)ESMFold (ESM-2 15B) (10GB Q4)
Recommended system RAM for AI: 16GB+ (2x GPU VRAM for model overflow)
Performance Estimates
Estimated tokens/sec for LLM inference based on 272 GB/s memory bandwidth — not hardware benchmarks. Methodology · What is Q4/Q8?
Llama 3.1 8B8B
Q8~20-24 tok/sUsableQwen 2.5 14B14B
Offload~1-3 tok/sVery slowMistral 7B7B
Q8~23-28 tok/sUsablePros
- +Affordable
- +Very low power
- +Good 1080p card
Cons
- -Only 8GB VRAM
- -Weak for AI workloads
- -128-bit bus
gamingbudget
Will It Run?
Llama 3.1 8B8B
Q8Qwen 2.5 14B14B
OffloadMistral 7B7B
Q8FLUX.1 Dev12B
OffloadStable Diffusion XL6.6B
Q8Stable Diffusion 3.5 Large8B
Q4CogVideoX-5B5B
Q4Mochi 110B
Offload