Catalog

Overview

🚀 LLM Ollama + OpenClaw on NVIDIA Jetson™

Overview

Stop just chatting with AI—start deploying agents that work for you. The LLM Ollama + OpenClaw container provides a streamlined, hardware-accelerated environment for developing and deploying autonomous AI applications on NVIDIA Jetson™. By integrating OpenClaw, this stack transforms standard LLMs into proactive assistants capable of executing tasks, managing files, and browsing the web across 20+ messaging channels (WhatsApp, Telegram, Slack, etc.).

Optimized for the edge, this container ensures high performance, low latency, and 100% private AI interaction, making it the ideal "Mission Control" for industrial, enterprise, and personal automation.


💻 Host System Requirements

To ensure optimal performance and hardware acceleration, your host system must meet the following specifications:

Component Version / Requirement
JetPack 6.2 (Production Release)
CUDA 12.6.68
cuDNN 9.3.0.75
TensorRT 10.3.0.30
OpenCV 4.8.0 (with CUDA support)

[!NOTE] Software versions (CUDA, TensorRT, etc.) are tied to JetPack 6.x. Please refer to the NVIDIA JetPack Documentation for compatibility details.


🌟 Key Features

Feature Description
Ollama Backend Run large language models (LLMs) locally with simple setup and high-speed inference.
OpenClaw Agent Beyond chat: A proactive assistant that executes commands and manages files via messaging apps.
20+ Channels Connect your AI to WhatsApp, Telegram, Discord, Slack, and more out-of-the-box.
ClawHub Ecosystem Access 700+ community-built skills (Gmail, GitHub, Calendar, IoT control).
Control UI Browser-based dashboard (Port 18789) for managing channels, models, and sessions.
Hardware Accelerated Fully optimized for NVIDIA Ampere architecture (Jetson Orin series).
Offline-First Entirely self-hosted. Your data stays on-device with no internet required after setup.
Developer Ready Includes a REST API, Modelfile customization, and Docker-friendly architecture.

🏗️ Container Description

Quick Information

The build.sh script initializes the following environment:

Container Name Description
LLM-Ollama-OpenClaw-Jetson A unified, hardware-accelerated environment containing the complete AI stack.

Container Highlights

  • Ollama Engine: Serves as the "Local Brain," managing model quantization (.gguf) and streaming output.
  • OpenClaw Gateway: Runs as a single Node.js process to route messages, maintain persistent memory across sessions, and invoke tools.
  • Skill Integration: Allows the agent to use "eyes and hands" to interact with the local filesystem, terminal, and web browsers.

🚀 Use Cases: From Chatting to Doing

Use Case Transformation: The Agentic Advantage
Proactive AI Assistant Execute tasks, not just prompts. Send a message via Telegram to run scripts, summarize local logs, or move files.
Industrial Edge AI Use Ollama as the reasoning core to parse commands and make decisions based on local sensor data.
Private RAG Systems Query sensitive internal manuals or technical docs offline using a vector database integration.
Omnichannel Automation Use ClawHub skills to bridge your edge device with external tools like GitHub, Jira, or Google Calendar.
Multilingual Support Deploy localized assistants that translate and summarize in real-time without cloud latency or costs.

📦 Model Information

The container is configured to use Ollama with OpenClaw for reasoning. The default model and behavior can be adjusted via the .env file.

Item Description
Default Model MODEL_NAME in .env (Default: kimi-k2.5:cloud)
Model Source Ollama Model Library
Pull Command ollama pull <model-name>
Agent Link openclaw models set ollama/<model-name>

🛠️ Software Components

Component Version Description
Ollama 0.5.7 High-performance LLM inference engine.
OpenClaw Latest The agentic gateway and multi-channel assistant.
PyTorch 2.0.0+ Deep learning framework for local development.
ONNX Runtime 1.16.3 Cross-platform inference engine support.

🏁 Quick Start

For a detailed walkthrough, including the docker-compose.yml file and advanced networking setup, please visit the:

Advantech Container GitHub Repository

⚡ Installation

  1. Ensure Docker, Docker Compose, and NVIDIA Container Toolkit are installed.
  2. Clone the repository and run the build script:
git clone <repository-url>
cd LLM-Ollama-OpenClaw-on-NVIDIA-Jetson
chmod +x build.sh && sudo ./build.sh


💡 Best Practices & Recommendations

Model Selection: For Jetson Orin Nano/NX (8GB), we recommend models ≤ 3B parameters with Q4/Q8 quantization.

GPU Offloading: Ensure models are 100% loaded into GPU memory (VRAM) for zero-latency interactions.

Jetson Clocks: Run sudo jetson_clocks to lock frequencies for maximum inference stability.

Context Window: OpenClaw performs best with a context window of 64k tokens or higher for complex tool-calling.

⚠️ Known Limitations

First-Run Setup: Internet access is required during the initial build to install OpenClaw via npm.

RAM Usage: The container occupies ~5GB RAM on Orin NX 8GB. For smaller devices, increasing the NVMe swap size is highly recommended.

Loading Delay: Initial model inference may take ~10 seconds as the model weights are loaded into the GPU.

Copyright © 2026 Advantech Corporation. All rights reserved.