GPU Memory Estimator

Estimate GPU memory requirements for AI models.

Model Configuration

B

Total GPU Memory Required

13.6 GB

For inference

🧠Model Weights
13.0 GB
💾KV Cache
0.5 GB

Memory Breakdown

Model Weights13.04 GB
KV Cache0.52 GB

Recommended GPUs

RTX 3090 (24GB)RTX 4080 (16GB)RTX 4090 (24GB)A10 (24GB)A100 40GB (40GB)A100 80GB (80GB)H100 (80GB)