GPU Memory Estimator
Estimate GPU memory requirements for AI models.
Model Configuration
B
Total GPU Memory Required
13.6 GB
For inference
π§ Model Weights
13.0 GB
πΎKV Cache
0.5 GB
Memory Breakdown
Model Weights13.04 GB
KV Cache0.52 GB
Recommended GPUs
RTX 3090 (24GB)RTX 4080 (16GB)RTX 4090 (24GB)A10 (24GB)A100 40GB (40GB)A100 80GB (80GB)H100 (80GB)