Quickly Deploy DeepSeek on reComputer Jetson
Introduction
DeepSeek is a cutting-edge AI model suite optimized for efficiency, accuracy, and real-time processing. With advanced optimization for edge computing, DeepSeek enables fast, low-latency AI inference directly on Jetson devices, reducing dependency on cloud computing while maximizing performance.

This wiki provides a step-by-step guide to deploying DeepSeek models on reComputer Jetson devices for efficient AI inference on the edge.
Prerequisites
- Jetson device with more than 8GB of memory.
- The jetson device needs to be pre-flashed with the jetpack 5.1.1 operating system or later.
In this wiki, we will accomplish the following tasks using the reComputer J4012 - Edge AI Computer with NVIDIA® Jetson™ Orin™ NX 16GB, but you can also try using other Jetson devices.

Getting Started
Hardware Connection
- Connect the Jetson device to the network, mouse, keyboard, and monitor.
Of course, you can also remotely access the Jetson device via SSH over the local network.
Install Ollama Inference Engine
Ollama is a lightweight and efficient inference engine designed for running large language models (LLMs) locally with minimal setup. It simplifies the deployment of AI models by providing an easy-to-use interface and optimized runtime for various hardware configurations, including Jetson devices.
To install Ollama, open the terminal window on the Jetson device and run the following command:
curl -fsSL https://ollama.com/install.sh | sh

This script will automatically download and set up Ollama on your system, enabling seamless local inference for AI applications.
Load and Run DeepSeek
Ollama now supports various versions of the DeepSeek models, allowing us to deploy different model sizes based on our needs. For demonstration purposes, we will use the default DeepSeek-R1 7B model.

ollama run deepseek-r1
This command downloads and prepares the DeepSeek model for local inference using Ollama. Once the model has finished loading, you can enter your query in the terminal window.

Effect Demonstration
In the demonstration video, the Jetson device operates at just 20W yet achieves an impressive inference speed.
References
- https://www.deepseek.com/
- https://ollama.com/library/deepseek-r1
- https://wiki.seeedstudio.com/local_ai_ssistant/
- https://www.seeedstudio.com/tag/nvidia.html
Tech Support & Product Discussion
Thank you for choosing our products! We are here to provide you with different support to ensure that your experience with our products is as smooth as possible. We offer several communication channels to cater to different preferences and needs.