DeepSeek-R1 model

Demo Environment

Development board:Jetson Orin Series Development Board

SSD:256G

Tutorial application scope: Whether the motherboard can run is related to the available memory of the system. The user's own environment and the programs running in the background may cause the model to fail to run.

Development board modelRun directly with OllamaRun with Open WebUI
Jetson Orin NX 16GB
Jetson Orin NX 8GB
Jetson Orin Nano 8GB
Jetson Orin Nano 4GB√(need to run the small parameter version)√(need to run the small parameter version)

DeepSeek-R1 is an open source large language model (LLM) designed by DeepSeek to understand and generate code.

Model scale

ModelParameters
DeepSeek-r11.5B
DeepSeek-r17B
DeepSeek-r18B
DeepSeek-r114B

 

Pull DeepSeek-r1

When using the large language model image provided by Yabo Intelligence, please note that the DeepSeek-R1 model is not pre-installed in the image.

The pull command will automatically pull the model from the Ollama model library. For orin NX 16Gb, you can run the following command to pull the model:

When pulling a model with large parameters such as DeepSeek-R1, it is recommended that you check the system's available memory in advance to ensure that the SSD system has completed the memory expansion operation to avoid deployment failures or other problems caused by insufficient memory.

orin 8Gb motherboard can run the following command to pull the model

Small parameter version model: orin 4Gb motherboard with running memory can run this

Using DeepSeek-r1

Run DeepSeek-r1

When using the large language model image provided by Yabo Intelligence, please note that the DeepSeek-R1 model is not pre-installed in the image.

When pulling or running a model with large parameters such as DeepSeek-R1, it is recommended that you check the system's available memory in advance to ensure that the SSD solid-state system has completed the memory expansion operation to avoid deployment failures or other problems caused by insufficient memory

orin nx super 16gb mainly runs the following parameter models

The development board with 8G RAM can run this:

A motherboard with 4G RAM can run this:

Have a conversation

The response time is related to the hardware configuration, please be patient!

orin nx super 16GB test:

Startup instructions:

orin super 8GB test:

Startup Instructions:

orin super 4GB test:

Startup instructions:

If the terminal does not reply to the message during the conversation, it means that the ollama version is too low. You can run the following command to update the ollama version

Run the update command

 

End the conversation

Use the Ctrl+d shortcut or /bye to end the conversation!

 

References

Ollama

Official website:https://ollama.com/

GitHub:https://github.com/ollama/ollama

DeepSeek-r1

Ollama corresponding model::https://ollama.com/library/deepseek-r1

GitHub:https://github.com/deepseek-ai/DeepSeek-r1