Skip to content

A simple and easy-to-use GPU memory calculator for Large Language Models (LLMs). Helps you quickly estimate the GPU memory requirements and number of devices needed to run models of various sizes.一个简单易用的大型语言模型 (LLM) GPU 内存需求计算器。帮助你快速估算运行不同规模模型所需的 GPU 内存和设备数量。

Notifications You must be signed in to change notification settings

cloud26/llm-gpu-memory-calculator

Repository files navigation

LLM GPU Memory Calculator

A simple and easy-to-use GPU memory calculator for Large Language Models (LLMs). Helps you quickly estimate the GPU memory requirements and number of devices needed to run models of various sizes.

English | 简体中文

Features

  • Calculate required GPU memory based on model parameter count
  • Support for multiple precision formats: FP32, FP16/BFLOAT16, FP8, INT8, INT4
  • Built-in presets for popular LLM models (DeepSeek-R1, Qwen, Llama, etc.)
  • Support for over 130 GPU models, including:
    • NVIDIA Data Center GPUs (H100, H200, B100, B200, etc.)
    • NVIDIA Consumer GPUs (RTX series)
    • AMD Data Center GPUs (Instinct series)
    • AMD Consumer GPUs (RX series)
    • Apple Silicon (M1-M4 series)
    • Huawei Ascend series
  • Complete internationalization support (English, Simplified Chinese, Traditional Chinese, Russian, Japanese, Korean, Arabic)
  • Responsive design for desktop and mobile devices

Calculation Method

This calculator uses the following formula to estimate GPU memory required for LLM inference:

  1. Model Weight Memory = Number of Parameters × Bytes per Parameter
  2. Inference Memory = Model Weight Memory × 10% (for activations, KV cache, etc.)
  3. Total Memory Requirement = Model Weight Memory + Inference Memory
  4. Required GPUs = Total Memory Requirement ÷ Single GPU Memory Capacity (rounded up)

Tech Stack

Local Development

# Install dependencies
pnpm install

# Start development server
pnpm dev

# Build for production
pnpm build

# Start production server
pnpm start

Contributing

Contributions via Pull Requests or Issues are welcome to help improve this project. Potential areas for contribution include:

  • Adding more GPU models
  • Adding more LLM model presets
  • Optimizing calculation methods
  • Adding support for more languages
  • Improving UI/UX

License

MIT

About

A simple and easy-to-use GPU memory calculator for Large Language Models (LLMs). Helps you quickly estimate the GPU memory requirements and number of devices needed to run models of various sizes.一个简单易用的大型语言模型 (LLM) GPU 内存需求计算器。帮助你快速估算运行不同规模模型所需的 GPU 内存和设备数量。

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published