DeepSeek’s first-generation reasoning models, achieving performance comparable to OpenAI-o1 across math, code, and reasoning tasks.
Distilled models
DeepSeek team has demonstrated that the reasoning patterns of larger models can be distilled into smaller models, resulting in better performance compared to the reasoning patterns discovered through RL on small models.
These models are created via fine-tuning against several dense models widely used in the research community using reasoning data generated by DeepSeek-R1. The evaluation results demonstrate that the distilled smaller dense models perform exceptionally well on benchmarks.
Ollama is a powerful tool designed to simplify the interaction with Large Language Models (LLMs). It allows you to easily download, manage, and deploy various LLMs for tasks like natural language processing, chatbots, and content generation.
Package | Version | License |
---|---|---|
DeepSeek | R1-Distill-Llama-8B | MIT License |
Click the Deploy to DigitalOcean button to create a Droplet based on this 1-Click App. If you aren’t logged in, this link will prompt you to log in with your DigitalOcean account.
In addition to creating a Droplet from the DeepSeek R1 Distill Llama 8B - Multi GPU 1-Click App using the control panel, you can also use the DigitalOcean API. As an example, to create a 4GB DeepSeek R1 Distill Llama 8B - Multi GPU Droplet in the SFO2 region, you can use the following curl
command. You need to either save your API access token) to an environment variable or substitute it in the command below.
curl -X POST -H 'Content-Type: application/json' \
-H 'Authorization: Bearer '$TOKEN'' -d \
'{"name":"choose_a_name","region":"sfo2","size":"s-2vcpu-4gb","image": "digitaloceanai-deepseekr1distillllama8b8x"}' \
"https://api.digitalocean.com/v2/droplets"
Access the Droplet Console:
root
:ssh root@your_droplet_public_IP
+ Ensure your SSH key is added to the SSH agent, or specify the key file directly:
ssh -i /path/to/your/private_key root@your_droplet_public_IP
+ Once connected, you will be logged in as the root user without needing a password.
Check the Message of the Day (MOTD) for Access Token:
Use your droplet IP and Access Token to start making requests to the API
curl --location 'http://<*droplet_ip*>/api/generate' \
--header 'Content-Type: application/json' \
--header 'Authorization: Bearer <*your_token_here*>' \
--data '{
"model": "deepseek-r1:8b",
"prompt": "What is DeepSeek-R1?",
"stream": false
}'
For more information please refer to Ollama API docs
sudo systemctl status caddy
Ollama is configured to run as systemd services for easy management. You can manage these services using the following commands:
systemctl status ollama
To secure your with HTTPS, you can configure TLS using Certbot and Caddy.
sudo apt-get update
sudo apt-get install certbot
Run the following command to obtain a free SSL certificate from Let’s Encrypt:
sudo certbot certonly --standalone -d <your_domain>
Replace your_domain
with your actual domain name.
Edit the Caddy configuration file located at /etc/caddy/caddyfile
. Update it to include the following settings:
:443 {
tls /etc/letsencrypt/live/<your_domain>/fullchain.pem /etc/letsencrypt/live/<your_domain>/privkey.pem
reverse_proxy localhost:8080
log {
output file /var/log/caddy/access.log
}
}
Replace your_domain
with your actual domain name.
After making changes to the Caddyfile, restart the Caddy service to apply the new configuration:
sudo systemctl restart caddy
Fail2Ban is configured to provide additional security by monitoring login attempts and banning IP addresses that show malicious signs. The rules for Open WebUI are defined as follows: