Run DeepSeek R1 locally on your device (Beginner-Friendly Guide)

DeepSeek R1 is one of the best open-source models in the market right now, and you can run DeepSeek R1 on your own computer!
DeepSeek R1 requires data-center level computers to run at its full potential, and weâll use a smaller version that works great on regular computers.
Why use an optimized version?
- Efficient performance on standard hardware
- Faster download and initialization
- Optimized storage requirements
- Maintains most of the original modelâs capabilities
Quick Steps at a Glance
- Download Jan
- Select a model version
- Choose settings
- Set up the prompt template & start using DeepSeek R1
Letâs walk through each step with detailed instructions.
Step 1: Download Jan
Jan is an open-source application that enables you to run AI models locally. Itâs available for Windows, Mac, and Linux. For beginners, Jan is the best choice to get started.

- Visit jan.ai
- Download the appropriate version for your operating system
- Install the app
Step 2: Choose Your DeepSeek R1 Version
To run AI models like DeepSeek R1 on your computer, youâll need something called VRAM (Video Memory). Think of VRAM as your computerâs special memory for handling complex tasks like gaming or, in our case, running AI models. Itâs different from regular RAM - VRAM is part of your graphics card (GPU).
Letâs first check how much VRAM your computer has. Donât worry if itâs not much - DeepSeek R1 has versions for all kinds of computers!
Finding your VRAM is simple:
- On Windows: Press
Windows + R
, typedxdiag
, hit Enter, and look under the âDisplayâ tab - On Mac: Click the Apple menu, select âAbout This Macâ, then âMore Infoâ, and check under âGraphics/Displaysâ
- On Linux: Open Terminal and type
nvidia-smi
for NVIDIA GPUs, orlspci -v | grep -i vga
for other graphics cards
Once you know your VRAM, hereâs what version of DeepSeek R1 will work best for you. If you have:
- 6GB VRAM: Go for the 1.5B version - itâs fast and efficient
- 8GB VRAM: You can run the 7B or 8B versions, which offer great capabilities
- 16GB or more VRAM: You have access to the larger models with enhanced features
Available versions and basic requirements for DeepSeek R1 distills:
Version | Model Link | Required VRAM |
---|---|---|
Qwen 1.5B | DeepSeek-R1-Distill-Qwen-1.5B-GGUF | 6GB+ |
Qwen 7B | DeepSeek-R1-Distill-Qwen-7B-GGUF | 8GB+ |
Llama 8B | DeepSeek-R1-Distill-Llama-8B-GGUF | 8GB+ |
Qwen 14B | DeepSeek-R1-Distill-Qwen-14B-GGUF | 16GB+ |
Qwen 32B | DeepSeek-R1-Distill-Qwen-32B-GGUF | 16GB+ |
Llama 70B | DeepSeek-R1-Distill-Llama-70B-GGUF | 48GB+ |
To download your chosen model:
Launch Jan and navigate to Jan Hub using the sidebar

- Input the model link in this field:

Step 3: Configure Model Settings
When configuring your model, youâll encounter quantization options:
Step 4: Configure Prompt Template
Final configuration step:
- Access Model Settings via the sidebar
- Locate the Prompt Template configuration
- Use this specific format:
This template is for proper communication between you and the model.
Youâre now ready to interact with DeepSeek R1:
