Can NVIDIA GeForce RTX 4060 Ti 8GB run FLUX.1 Dev?

12B parameter Image Gen model on 8GB GDDR6

Barely — requires CPU/RAM offloading
~1-3 tok/s (offload)
SpeedVery slow — expect 1-3 tokens/sec
QualityQuality is fine but the speed makes it impractical for interactive use

VRAM Requirements

FLUX.1 Dev is a 12B parameter model. At full precision (FP16), it requires 32GB of VRAM. Your NVIDIA GeForce RTX 4060 Ti 8GB only has 8GB — not enough even at maximum compression.

FP16 (Full Precision)32GB (need 24GB more)

Maximum quality, no quantization

Q8 (8-bit)16GB (need 8GB more)

Near-lossless, ~50% size reduction

Q4 (4-bit)10GB (need 2GB more)

Good quality, ~75% size reduction

Your GPU VRAM: 8GB GDDR6 at 288 GB/s bandwidth
Recommended system RAM: 32GB DDR5 (2x GPU VRAM minimum for model overflow)

How to Set It Up

Step 1: Install ComfyUI

git clone https://github.com/comfyanonymous/ComfyUI.git && cd ComfyUI && pip install -r requirements.txt

ComfyUI is the recommended UI for Stable Diffusion and FLUX models.

Step 2: Download the model

Download FLUX.1 Dev weights from HuggingFace and place them in ComfyUI/models/. The model is approximately 32GB at full precision.

Step 3: Launch and generate

python main.py

Open http://localhost:8188 in your browser. You can use the full precision weights.

NVIDIA GeForce RTX 4060 Ti 8GB Specs

VRAM8GB GDDR6
Memory Bandwidth288 GB/s
TDP160W
CUDA Cores4,352
Street Price~$370
AI Rating3/10

Other Image Gen Models on NVIDIA GeForce RTX 4060 Ti 8GB

About FLUX.1 Dev

State-of-the-art image generation. 16GB comfortable at FP8.

Category: Image Gen · Parameters: 12B · CUDA required: Recommended