Welcome to the Step-by-Step Guidance version of this project. Let's do this!
π£ If you're EVER stuck - ask the NextWork community. Students like you are already asking questions about this project.
Weβll start with the easiest way to use DeepSeek - its web app.
If youβve used ChatGPT or any other LLM in your browser, this will feel familiar. But, as we dive into more advanced prompts, youβll start noticing what makes DeepSeek stand out.
In this step, you're going to:
Create a DeepSeek Account
πββοΈ I don't want to enter my email
You could consider using a temporary email address instead.For example, you could use a tool like TempMail to set up a temporary email address that receives mail for 1-2 hours, and deletes itself right after. Just make sure to keep this tab open until the end of the project, so you don't lose the inbox.
πββοΈ I'm not getting a code
You might not get a code if DeepSeek's registration is busy. You could try signing up with Google instead.If Google isn't an option or isn't working either, you could skip this step for now and head to the next step. The first step is great for easy access to DeepSeek, but you can always come back to it later!
Run a Prompt
π‘ Tip: You can choose another prompt to give DeepSeek. We'd recommend a short and easy request (e.g. ask for a 100 word summary instead of a 1000 word essay), so you won't be waiting too long for an answer!
π‘ What is DeepThink (R1)?
DeepThink (R1) is DeepSeek's latest AI model. It stands out for displaying its real-time reasoning process before generating a response.Reasoning is a big deal in LLMs. Unlike older models that simply predict the next word in a sequence, DeepThink - as well as OpenAIβs o1 model - is one of the first models to actively reviews its own responses as it generates them.
In R1's thinking process, spot for intermediate thinking steps like self-doubt ("hmm") and verification checks ("wait"), which gives you a lot of transparency into its problem-solving approach. Reasoning also makes an LLM much more efficient - they're more likely to solve a problem in one go, without requiring lots of back and forth between you and the LLM.
Test DeepSeek vs ChatGPT on Advanced Reasoning
Let's challenge DeepSeek to a harder prompt that requires more reasoning, and see how it compares with another LLM (e.g. OpenAI).
Now that you have a feel of how DeepSeek works, let's see how we can host it locally without relying on the web app.
π‘ What are the downsides of using a web app?
Ooo good question! Web-based LLMs, like ChatGPT and DeepSeek online:
- Require constant internet connection (you can't go offline)
- Introduce latency (web apps run slower when there are lots of people sending requests)
- Process queries through external servers - which might raise privacy concerns around how the data is stored and used.
π‘ What does it mean to host DeepSeek locally?
Running DeepSeek on your own computer means you donβt need the web app at all. Your device does all the processing, so no external servers are involved.That means you can use DeepSeek offline, keep all your data private, and get faster responses since there's no waiting on the internet.
In this step, you're going to:
Download Ollama
Note: If you already have Ollama installed, you can skip ahead to the
next step .
π‘ What is Ollama?
Ollama is a tool that makes it easy to host LLMs, like DeepSeek, on your own computer. You can start chatting with LLMs over your computer's terminal!Ollama takes care of downloading, installing, and running the models, so you don't have to worry about the complex setup that comes with hosting an LLM locally.
Ollama also gives you more control around the LLM you're using. We'll experiment with a setting called temperature later in this project to see the benefits of having wider control.
Install Ollama
Next up, installing Ollama! Installation instructions depend on your operating system.
π‘ Haven't I already installed Ollama?
So far you've just downloaded Ollama's installation files, which means Ollama is like a package that's been delivered to your door - but you haven't opened the package yet.You'll need to open the package and set up permissions to start using Ollama's software in your computer.
Nice! Now Ollama will take you through the process of installing the software locally.
Enter the password you use to unlock your Mac. This gives Ollama the permission to install the software into your computer.
In the next panel, we'll ignore the command that Ollama gives us. The command lets us use another open-source LLM, but we'll go straight to using DeepSeek.
Open your
After downloading the executable file, simply run it, and Ollama will be installed automatically.
If you're stuck or need any troubleshooting, we'd recommend checking out the Windows instructions on Ollama's GitHub.
For the most up to date instructions, we'd recommend visiting Ollama's GitHub
Manual install
[!NOTE] If you are upgrading from a prior version, you should remove the old libraries with
sudo rm -rf /usr/lib/ollama first.
Download and extract the package:
Start Ollama:
In another terminal, verify that Ollama is running:
AMD GPU install
If you have an AMD GPU, also download and extract the additional ROCm package:
ARM64 install
Download and extract the ARM64-specific package:
Adding Ollama as a startup service (recommended)
Create a user and group for Ollama:
Create a service file in
Then start the service:
Now that we've installed Ollama, how do we use it to access DeepSeek locally?
In this step, you're going to:
Find and Install DeepSeek R1
π‘ Why can't I find OpenAI's models on Ollama?
Ollama focuses on open-source models like DeepSeek.OpenAI's models are closed systems, so the underlying architecture, codebase, and datasets used to develop OpenAI models are confidential. Because they're confidential, it's not possible to use OpenAI locally in your machine.
π‘ What are these different dropdown options?
The different dropdown options represent different model sizes for R1. Think of DeepSeek R1 in the web app as R1 at full capacity - if you wanted to run this version of R1 locally, you would need a computer with very large processing power and storage space (the dropdown tells us it requires 404GB of storage). This is far beyond what most computers can handle, as computers typically have less available storage and memory.
Model sizes let you choose a smaller, more accessible version of DeepSeek R1 for local use.
Smaller models (like 1.5b) are faster and require less memory to run locally, while larger models (like 8b) have deeper reasoning abilities and are more accurate. We're installing 1.5b first as a quick start, but we'll use a larger model next to the difference in performance.
π‘ What does "1.5b" mean?
In AI models like DeepSeek, "1.5 b" means the model has 1.5 billion parameters to learn patterns from data.Think of parameters as tiny decision-makers inside the model, each helping it recognize patterns, analyze data, and improve reasoning. More parameters generally mean the model can handle more complex tasks, but bigger isnβt always better - it also depends on how well the model is trained.
π‘ What does this command do?
This command sets up DeepSeek's smallest model, i.e. the 1.5 billion parameter model, locally in your computer. Because the command usesrun , your terminal will transform into a chat session with DeepSeek R1 too.
π‘ Extra for Experts: The terminal response starts by 'pulling manifest' - what does that mean?
When you run the Ollama command, it fetches the DeepSeek model's manifest, which is like a blueprint that tells your computer how to set up and run the model. It includes instructions for downloading and configuring everything correctly.
The actual brain of DeepSeek is the model itself, which gets downloaded after the manifest. Think of the manifest as the setup guide, while the model is the intelligence your computer will use to process prompts and generate responses.
Use Another DeepSeek R1 Model
πββοΈ How do I know how much storage my computer has?
Option 1: Settings
Option 2: File Explorer
Option 1: Terminal Command
Option 2: Disk Usage Analyzer
If you're stuck picking a model size, we'd recommend going for the 8b option. If there are any issues with using it, you can always switch to the 7b option instead.
Test Prompts
π‘ Why can I still access DeepSeek while offline?
Local hosting through Ollama means you don't need another server to process your prompt. The DeepSeek model is running entirely on your device without needing internet connectivity.
π‘ What are the<think> tags?
The<think> tags are a terminal version of DeepSeek's real-time reasoning display, so you can still see how DeepSeek is generating its response.You might've noticed that the the think tags were empty in your previous request. That's because
Hello was a more straightforward prompt, so deep thinking (which triggers this real-time reasoning display) wasn't required.
While the terminal is great for quick tests, you might miss the look of the web app. It does a much better job of organizing your chats and making conversations user friendly!
No worries, you can use a tool called Chatbox to organize your conversations in the terminal to look like the web app too. Let's set that up!
In this step, you're going to:
Install and Configure Chatbox
πββοΈ How do I find the correct option for my operating system?
π MacOS: Select the Apple icon from the top left hand corner of your computer's menu bar. Select About this Mac, and note whether your Chip says Apple (pick Apple Silicon) or Intel.
πΌοΈ Windows: Select the Start button and search for System Information. Note whether your System Type says x64-based or ARM-based PC.
π§ Linux: Open a terminal and run
uname -m . Match the output to the correct Linux package option.
π‘ What does the Model provider setting do?
In Chatbox, the Model provider determines the API that will connect you to the LLM model you want to use. We're using the Ollama API, since Ollama is the tool we're using to run DeepSeek locally.
π‘ Why are we leaving the API Host as the default?
To connect you with your local LLM, Ollama needs to set up an endpoint, which is like an address within your computer to run DeepSeek. Ollama sets up LLMs at a default location (127.0.0.1:11434), so we'll keep the default value in Chatbox. This setup lets Chatbox communicate directly with your locally-hosted DeepSeek model.
Chat with DeepSeek
πββοΈ DeepSeek made an error!
You might notice that the 1.5b model gave you an incorrect answer! Instead of three r's in strawberry, it only found two.The 1.5b model is the most lightweight R1 model, so it's less able to analyse text and conduct proper reasoning.
πββοΈ DeepSeek didn't make an error!
How good is that! It's great if DeepSeek got that right with a smaller model. A comparison where both models produce the right result is still a great experiment. You could always try other problems or prompts to test the limits π₯
Let's compare this with your larger local DeepSeek model!
Select another New Chat in the bottom left corner of Chatbox.
This time, let's switch the DeepSeek model from 1.5b to 8b.
Let's explore an advanced setting, called temperature, to see how you can customize DeepSeek R1 depending on the use case.
In this step, you're going to:
High Temperature Test
π‘ What is temperature?
Temperature controls the randomness of an LLM's output.A higher temperature, like 2, gives you more creative and unpredictable responses, while a lower temperature, like 0, gives more focused and logical responses.
This is a detailed setting that you might not have access to over a web app, but possible with local hosting and APIs. Chatbox makes it easy for you to edit and customise your AI model's temperature.
In this example, you might notice that the recipe uses orange juice! Interesting ingredient choice...
Select your Settings from the bottom left hand corner again.
Change the temperature to
Low Temperature Test
Set up a Third Chat
π‘ Why are we opening ChatGPT?
This new chat will act as the judge of the two responses we generate - can another AI tell the difference between high and low temperature responses?You'll also get to learn a breakdown of how to detect low vs high temperature text along the way.
Now head back to the DeepSeek chat where your second respoonse has been generating.
Your second response might take more time to generate, so you might need to wait a few more minutes.
Review DeepSeek's response - do you see any difference in the tone of voice and choice of words? π΅οΈββοΈ
Oooo, nice work ChatGPT. You might notice that ChatGPT can correctly point out the first response had a higher temperature setting, and why.
π‘ Why does temperature matter?
Different temperature settings work great for different scenarios.
- Translation services and code commenting usually use a low temperature setting (0.2-0.4) to prioritize accuracy.
- Chatbots usually use a low-mid temperature setting (0.5), so that responses are true and accurate.
- For summarising meetings or writing email responses, you might opt for a mid-high temperature setting (0.6-0.8) to add a bit of variety and generated insights.
- For creative writing and brainstorming, opt for a higher temperature setting (0.9-1.5) to get a wide range of results.
π‘ Extra for Experts: I want to try another temperature experiment! We got you! Challenge your DeepSeek model to generate two responses - one with high temperature (2), one with low temperature (0) - to the following prompt:Write a short 100 word story set in a world where gravity changes direction every day. See the differences between a creative high-temperature story, versus a logical low-temperature story!
Welcome to your π€« exclusive π€« secret mission! Are you ready for the ultimate test?
Your mission, should you choose to accept it, is to expose how efficiently DeepSeek and OpenAI use tokens. This lets you know which model gives you the most value. Letβs dive in!
In this secret mission, you're going to:
Now that we've explored the world of LLMs with DeepSeek and Ollama, it's time to clean up. This is important to keep your systems tidy.
Resources to delete:
Remove the DeepSeek models from Ollama (optional).
Uninstall Ollama (optional).
Uninstall Chatbox (optional).
If you no longer plan to use the DeepSeek models, you can remove them to free up disk space.
π MacOS
If you've installed Ollama via Homebrew, you can also uninstall it using
To check if Ollama is fully removed, run:
If it returns
πΌοΈ Windows
The Ollama Windows installer registers an Uninstaller application.
Under
π§ Linux
You've journeyed into the fascinating world of LLMs, and emerged victorious! π
You've learned how to:
π p.s. Does it say "Still tasks to complete!" at the bottom of the screen?
This means you still have screenshots left to upload, or questions left to answer!
- Press Ctrl+F (Windows) or Command+F (Mac) on your keyboard.
- Search for the text Return to later.
- Jump straight to your incomplete tasks!
- πββοΈ Still stuck? Ask the community!