Overview
About Advantech Container Catalog (ACC)
Advantech Container Catalog is a comprehensive collection of ready-to-use, containerized software packages designed to accelerate the development and deployment of Edge AI applications. By offering pre-integrated solutions optimized for embedded hardware, ACC simplifies the challenge of software-hardware compatibility, especially in GPU/NPU-accelerated environments.
Feature / Benefit | Description |
---|---|
Accelerated Edge AI Development | Ready-to-use containerized solutions for faster prototyping and deployment |
Hardware Compatible | Eliminates embedded hardware and software package incompatibility |
GPU/NPU Access Ready | Supports passthrough for efficient hardware acceleration |
Model Conversion & Optimization | Built-in AI model quantization and format conversion support |
Optimized for CV & LLM Applications | Pre-optimized containers for computer vision and large language models |
Container Overview
This container, GPU Passthrough on NVIDIA Jetson™, provides a ready-to-use environment with optimized AI frameworks, GPU passthrough, and industrial-grade reliability on NVIDIA Jetson platforms. It enables users to focus on developing AI applications on Advantech Edge AI systems accelerated by NVIDIA chipsets—eliminating the complexity of hardware setup and AI framework compatibility.
Key Features
- Full Hardware Acceleration: Optimized access to GPU, NVENC/NVDEC, and DLA
- Complete AI Framework Stack: PyTorch, TensorFlow, ONNX Runtime, and TensorRT™
- Industrial Vision Support: Accelerated OpenCV and GStreamer pipelines
- Edge AI Capabilities: Support for computer vision, LLMs, and time-series analysis
- Performance Optimized: Tuned specifically for Advantech EPC-R7300 and more devices
Host Device Prerequisites
Item | Specification |
---|---|
Compatible Hardware | Advantech devices accelerated by NVIDIA Jetson™ - refer to Compatible hardware |
NVIDIA Jetson™ Version | 5. x |
Host OS | Ubuntu 20.04 |
Required Software packages | *refer to below |
Software Installation | NVIDIA Jetson™ Software Package Installation |
Required Software Packages on Host Device
These packages are bound with NVIDIA Jetson™ version of the device. This container supports NVIDIA Jetson™ 5.x.
Component | Version | Description |
---|---|---|
CUDA® Toolkit | 11.4.315 | GPU computing platform |
cuDNN | 8.6.0.166 | Deep Neural Network library |
TensorRT™ | 8.5.2.2 | Inference optimizer and runtime |
VPI | 2.2.7 or above | |
Vulkan | 1.3.204 or above | |
OpenCV | 4.5.4 with CUDA®: NO |
Container Environment Overview
Software Components on Container Image
Component | Version | Description |
---|---|---|
CUDA® | 11.4.315 | GPU computing platform |
cuDNN | 8.6.0 | Deep Neural Network library |
TensorRT™ | 8.5.2.2 | Inference optimizer and runtime |
PyTorch | 2.0.0+nv23.02 | Deep learning framework |
TensorFlow | 2.12.0+nv23.05 | Machine learning framework |
ONNX Runtime | 1.16.3 | Cross-platform inference engine |
OpenCV | 4.5.0 | Computer vision library with CUDA® |
GStreamer | 1.16.2 | Multimedia framework |
Container Quick Start Guide
For container quick start, including docker-compose file, and more, please refer to Advantech EdgeSync Container Repository
Supported AI Capabilities
Vision Models
Model Family | Versions | Performance (FPS) | Quantization Support |
---|---|---|---|
YOLO | v3/v4/v5 (up to v5.6.0), v6 (up to v6.2), v7 (up to v7.0), v8 (up to v8.0) | YOLOv5s: 45-60 @ 640x640, YOLOv8n: 40-55 @ 640x640, YOLOv8s: 30-40 @ 640x640 | INT8, FP16, FP32 |
SSD | MobileNetV1/V2 SSD, EfficientDet-D0/D1 | MobileNetV2 SSD: 50-65 @ 300x300, EfficientDet-D0: 25-35 @ 512x512 | INT8, FP16, FP32 |
Faster R-CNN | ResNet50/ResNet101 backbones | ResNet50: 3-5 @ 1024x1024 | FP16, FP32 |
Segmentation | DeepLabV3+, UNet | DeepLabV3+ (MobileNetV2): 12-20 @ 512x512 | INT8, FP16, FP32 |
Classification | ResNet (18/50), MobileNet (V1/V2/V3), EfficientNet (B0-B2) | ResNet18: 120-150 @ 224x224, MobileNetV2: 180-210 @ 224x224 | INT8, FP16, FP32 |
Pose Estimation | PoseNet, HRNet (up to W18) | PoseNet: 15-25 @ 256x256 | FP16, FP32 |
Language Models Recommendation
Model Family | Versions | Memory Requirements | Performance Notes |
---|---|---|---|
DeepSeek Coder | Mini (1.3B), Light (1.5B) | 2-3 GB | 10-15 tokens/sec in FP16 |
TinyLlama | 1.1B | 2 GB | 8-12 tokens/sec in FP16 |
Phi | Phi-1.5 (1.3B), Phi-2 (2.7B) | 1.5-3 GB | Phi-1.5: 8-12 tokens/sec in FP16, Phi-2: 4-8 tokens/sec in FP16 |
Llama 2 | 7B (Quantized to 4-bit) | 3-4 GB | 1-2 tokens/sec in INT4/INT8 |
Mistral | 7B (Quantized to 4-bit) | 3-4 GB | 1-2 tokens/sec in INT4/INT8 |
DeepSeek R1 1.5B Optimizations Recommendations:
- Supports INT4-8 quantization for inference
- Best performance with TensorRT™ engine conversion
- Typical throughput: 8-12 tokens/sec in FP16, 12-18 tokens/sec in INT8
- Recommended batch size: 1-2 for real-time applications
Supported AI Model Formats
Format | Support Level | Compatible Versions | Notes |
---|---|---|---|
ONNX | Full | 1.10.0 - 1.16.3 | Recommended for cross-framework compatibility |
TensorRT™ | Full | 7.x - 8.5.x | Best for performance-critical applications |
PyTorch (JIT) | Full | 1.8.0 - 2.0.0 | Native support via TorchScript |
TensorFlow SavedModel | Full | 2.8.0 - 2.12.0 | Recommended TF deployment format |
TFLite | Partial | Up to 2.12.0 | May have limited hardware acceleration |
Hardware Acceleration Support
Accelerator | Support Level | Compatible Libraries | Notes |
---|---|---|---|
CUDA® | Full | PyTorch, TensorFlow, OpenCV, ONNX Runtime | Primary acceleration method |
TensorRT™ | Full | ONNX, TensorFlow, PyTorch (via export) | Recommended for inference optimization |
cuDNN | Full | PyTorch, TensorFlow | Accelerates deep learning primitives |
NVDEC | Full | GStreamer, FFmpeg | Hardware video decoding |
NVENC | Full | GStreamer, FFmpeg | Hardware video encoding |
DLA | Partial | TensorRT™ | Requires specific model optimization |
Copyright © 2025 Advantech Corporation. All rights reserved.