To run DeepSeek on Windows with a WebUI, you need to install Ollama, text-generation-webui, or another UI like Gradio. Below is the hardware requirement table for all model sizes.
DeepSeek WebUI Hardware Requirements
Model | VRAM (GPU) | RAM (System) | CPU | Storage (SSD/NVMe) | Recommended GPU |
---|---|---|---|---|---|
1.5B | 4GB+ | 16GB | Intel i5 / Ryzen 5 | 50GB | NVIDIA RTX 2060 |
7B | 16GB+ | 32GB | Intel i7 / Ryzen 7 | 100GB | NVIDIA RTX 3090 / 4090 |
8B | 24GB+ | 64GB | Intel i9 / Ryzen 9 | 150GB | NVIDIA RTX 4090 / A100 |
14B | 32GB+ | 128GB | Intel i9 / Ryzen 9 | 200GB | NVIDIA A100 (40GB) |
32B | 48GB+ | 256GB | AMD EPYC / Xeon | 400GB | NVIDIA H100 (80GB) |
70B | 80GB+ | 512GB | AMD EPYC / Xeon | 1TB | 2× NVIDIA H100 (80GB) |
671B | 512GB+ (Multiple GPUs) | 1.5TB+ | AMD EPYC / Xeon | 10TB+ | 8× NVIDIA H100 (80GB) |
Installation Steps for DeepSeek with WebUI on Windows
Option 1: Using text-generation-webui
- Install Dependencies
- Install Python 3.10+
- Install CUDA Toolkit 11.8+ (for NVIDIA GPUs)
- Install Git
- Download
text-generation-webui
git clone https://github.com/oobabooga/text-generation-webui.git
cd text-generation-webui
- Create Virtual Environment & Install Dependencies
python -m venv venv
venv\Scripts\activate
pip install -r requirements.txt
- Download a DeepSeek Model (Example: 7B)
- Download DeepSeek-7B GGUF model from Hugging Face:
https://huggingface.co/ - Place the downloaded model in
text-generation-webui/models/deepseek-7b
- Launch WebUI
python server.py --model deepseek-7b
- Open http://localhost:5000 in your browser.
Option 2: Using Ollama
Ollama provides an easy way to run DeepSeek with WebUI.
- Install Ollama: https://ollama.com
- Download & Run DeepSeek Model
ollama run deepseek
- Access WebUI:
- Go to http://localhost:11434
Notes:
- 1.5B – 8B models can run on high-end gaming GPUs (RTX 3090/4090).
- 14B+ models need professional AI hardware (A100, H100).
- 32B+ models require multiple GPUs with tensor parallelism.
- 671B model is too large for local use (requires cloud clusters).