Skip to content

Commit

Permalink
docs: add DeepSeek R1 local installation guide
Browse files Browse the repository at this point in the history
- Add comprehensive guide for running DeepSeek R1 locally
- Include step-by-step instructions with screenshots
- Add VRAM requirements and model selection guide
- Include system prompt setup instructions
  • Loading branch information
eckartal committed Jan 31, 2025
1 parent 3037206 commit 705116c
Show file tree
Hide file tree
Showing 9 changed files with 109 additions and 0 deletions.
Binary file added docs/src/pages/post/_assets/download-jan.jpg
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
109 changes: 109 additions & 0 deletions docs/src/pages/post/deepseek-r1-locally.mdx
Original file line number Diff line number Diff line change
@@ -0,0 +1,109 @@
---
title: "Beginner's Guide: Run DeepSeek R1 Locally (Private)"
description: "Quick steps on how to run DeepSeek R1 locally for full privacy. Perfect for beginners—no coding required."
tags: DeepSeek, R1, local AI, Jan, GGUF, Qwen, Llama
categories: guides
date: 2024-01-31
ogImage: assets/run-deepseek-r1-locally-in-jan.jpg
---

import { Callout } from 'nextra/components'
import CTABlog from '@/components/Blog/CTA'

# Beginner’s Guide: Run DeepSeek R1 Locally

![image](./_assets/run-deepseek-r1-locally-in-jan.jpg)

You can run DeepSeek R1 on your own computer! While the full model needs very powerful hardware, we'll use a smaller version that works great on regular computers.

Why use a smaller version?
- Works smoothly on most modern computers
- Downloads much faster
- Uses less storage space on your computer

## Quick Steps at a Glance
1. Download and install [Jan](https://jan.ai/) (just like any other app!)
2. Pick a version that fits your computer
3. Choose the best settings
4. Set up a quick template & start chatting!

Keep reading for a step-by-step guide with pictures.

## Step 1: Download Jan
[Jan](https://jan.ai/) is a free app that helps you run AI models on your computer. It works on Windows, Mac, and Linux, and it's super easy to use - no coding needed!

![image](./_assets/download-jan.jpg)

- Get Jan from [jan.ai](https://jan.ai)
- Install it like you would any other app
- That's it! Jan takes care of all the technical stuff for you

## Step 2: Choose Your DeepSeek R1 Version
DeepSeek R1 comes in different sizes. Let's help you pick the right one for your computer.

<Callout type="info">
💡 Not sure how much VRAM your computer has?
- Windows: Press Windows + R, type "dxdiag", press Enter, and click the "Display" tab
- Mac: Click Apple menu > About This Mac > More Info > Graphics/Displays
</Callout>

Below is a detailed table showing which version you can run based on your computer's VRAM:

| Version | Link to Paste into Jan Hub | Required VRAM for smooth performance |
|---------|---------------------------|---------------|
| Qwen 1.5B | [https://huggingface.co/bartowski/DeepSeek-R1-Distill-Qwen-1.5B-GGUF](https://huggingface.co/bartowski/DeepSeek-R1-Distill-Qwen-1.5B-GGUF) | 6GB+ VRAM |
| Qwen 7B | [https://huggingface.co/bartowski/DeepSeek-R1-Distill-Qwen-7B-GGUF](https://huggingface.co/bartowski/DeepSeek-R1-Distill-Qwen-7B-GGUF) | 8GB+ VRAM |
| Llama 8B | [https://huggingface.co/unsloth/DeepSeek-R1-Distill-Llama-8B-GGUF](https://huggingface.co/unsloth/DeepSeek-R1-Distill-Llama-8B-GGUF) | 8GB+ VRAM |
| Qwen 14B | [https://huggingface.co/bartowski/DeepSeek-R1-Distill-Qwen-14B-GGUF](https://huggingface.co/bartowski/DeepSeek-R1-Distill-Qwen-14B-GGUF) | 16GB+ VRAM |
| Qwen 32B | [https://huggingface.co/bartowski/DeepSeek-R1-Distill-Qwen-32B-GGUF](https://huggingface.co/bartowski/DeepSeek-R1-Distill-Qwen-32B-GGUF) | 16GB+ VRAM |
| Llama 70B | [https://huggingface.co/unsloth/DeepSeek-R1-Distill-Llama-70B-GGUF](https://huggingface.co/unsloth/DeepSeek-R1-Distill-Llama-70B-GGUF) | 48GB+ VRAM |

<Callout type="info">
Quick Guide:
- 6GB VRAM? Start with the 1.5B version - it's fast and works great!
- 8GB VRAM? Try the 7B or 8B versions - good balance of speed and smarts
- 16GB+ VRAM? You can run the larger versions for even better results
</Callout>

Ready to download? Here's how:
1. Open Jan and click the button in the left sidebar to open Jan Hub
2. Find the "Add Model" section (shown below)

![image](./_assets/jan-library-deepseek-r1.jpg)

3. Copy the link for your chosen version and paste it here:

![image](./_assets/jan-hub-deepseek-r1.jpg)

## Step 3: Choose Model Settings
When adding your model, you'll see two options:

<Callout type="tip">
- **Q4:** Perfect for most users - fast and works great! ✨ (Recommended)
- **Q8:** Slightly more accurate but needs more powerful hardware
</Callout>

## Step 4: Set Up & Start Chatting
Almost done! Just one quick setup:

1. Click Model Settings in the sidebar
2. Find the Prompt Template box
3. Copy and paste this exactly:

<Callout type="warning">
```
<|User|>{prompt}<|Assistant|>
```
</Callout>

This helps DeepSeek understand when you're talking and when it should respond.

Now you're ready to start chatting!

![image](./_assets/jan-runs-deepseek-r1-distills.jpg)

## Need help?

<Callout type="info">
Having trouble? We're here to help! [Join our Discord community](https://discord.gg/Exe46xPMbK) for support.
</Callout>

0 comments on commit 705116c

Please sign in to comment.