GPU Memory Estimator

Estimate GPU memory requirements for AI models.

Model Configuration

B

Total GPU Memory Required

13.6 GB

For inference

🧠Model Weights
13.0 GB
πŸ’ΎKV Cache
0.5 GB

Memory Breakdown

Model Weights13.04 GB
KV Cache0.52 GB

Recommended GPUs

RTX 3090 (24GB)RTX 4080 (16GB)RTX 4090 (24GB)A10 (24GB)A100 40GB (40GB)A100 80GB (80GB)H100 (80GB)