Overview
Container GPU Passthrough
About Advantech Container Catalog
Advantech Container Catalog is a comprehensive collection of ready-to-use, containerized software packages designed to accelerate the development and deployment of Edge AI applications. By offering pre-integrated solutions optimized for embedded hardware, it simplifies the challenges often faced with software and hardware compatibility, especially in GPU/NPU-accelerated environments.
Key benefits of the Container Catalog include:
Feature / Benefit | Description |
---|---|
Accelerated Edge AI Development | Ready-to-use containerized solutions for fast prototyping and deployment |
Hardware Compatibility Solved | Eliminates embedded hardware and AI software package incompatibility |
GPU/NPU Access Ready | Supports passthrough for efficient hardware acceleration |
Model Conversion & Optimization | Built-in AI model quantization and format conversion support |
Optimized for CV & LLM Applications | Pre-optimized containers for computer vision and large language models |
Scalable Device Management | Supports large-scale IoT deployments via EdgeSync, Kubernetes, etc. |
Lower Entry Barrier for Developers | High-level language (Python, C#, etc.) support enables easier development |
Developer Accessibility | Junior engineers can build embedded AI applications more easily |
Increased Customer Stickiness | Simplified tools lead to higher adoption and retention |
Open Ecosystem | 3rd-party developers can integrate new apps to expand the platform |
Container Overview
This container, Container GPU Passthrough, provides a ready-to-use environment with optimized AI frameworks, GPU passthrough, and industrial-grade reliability on Advantech hardware platforms accelerated by GPU. It enables users to focus on developing AI applications on Advantech Edge AI systems accelerated by GPU chipsets—eliminating the complexity of hardware setup and AI framework compatibility.
Key Features
- Full Hardware Acceleration: Optimized access to GPU
- Complete AI Framework Stack: PyTorch, TensorFlow, ONNX Runtime
- Industrial Vision Support: Accelerated OpenCV pipelines
- Edge AI Capabilities: Support for computer vision, LLMs, and time-series analysis
- Performance Optimized: Tuned specifically for Advantech EPC-R7300 and more devices
Host Device Prerequisites
Item | Specification |
---|---|
Compatible Hardware | Advantech GPU-accelerated devices - refer to Compatible hardware |
Host OS | Ubuntu 20.04 |
Required Software packages | *refer to below |
Software Installation | Host Software Package Installation |
Container Environment Overview
Container Quick Start Guide
For Software Components on Container Image, container quick start, including docker-compose file, and more, please refer to Advantech EdgeSync Container Repository
Supported AI Capabilities
Vision Models
Model Family | Versions | Performance (FPS) | Quantization Support |
---|---|---|---|
YOLO | v3/v4/v5 (up to v5.6.0), v6 (up to v6.2), v7 (up to v7.0), v8 (up to v8.0) | YOLOv5s: 45-60 @ 640x640, YOLOv8n: 40-55 @ 640x640, YOLOv8s: 30-40 @ 640x640 | INT8, FP16, FP32 |
SSD | MobileNetV1/V2 SSD, EfficientDet-D0/D1 | MobileNetV2 SSD: 50-65 @ 300x300, EfficientDet-D0: 25-35 @ 512x512 | INT8, FP16, FP32 |
Faster R-CNN | ResNet50/ResNet101 backbones | ResNet50: 3-5 @ 1024x1024 | FP16, FP32 |
Segmentation | DeepLabV3+, UNet | DeepLabV3+ (MobileNetV2): 12-20 @ 512x512 | INT8, FP16, FP32 |
Classification | ResNet (18/50), MobileNet (V1/V2/V3), EfficientNet (B0-B2) | ResNet18: 120-150 @ 224x224, MobileNetV2: 180-210 @ 224x224 | INT8, FP16, FP32 |
Pose Estimation | PoseNet, HRNet (up to W18) | PoseNet: 15-25 @ 256x256 | FP16, FP32 |
Supported AI Model Formats
Format | Support Level | Compatible Versions | Notes |
---|---|---|---|
ONNX | Full | 1.10.0 - 1.16.3 | Recommended for cross-framework compatibility |
PyTorch (JIT) | Full | 1.8.0 - 2.0.0 | Native support via TorchScript |
TensorFlow SavedModel | Full | 2.8.0 - 2.12.0 | Recommended TF deployment format |
TFLite | Partial | Up to 2.12.0 | May have limited hardware acceleration |
Language Models Recommendation
Model Family | Versions | Memory Requirements | Performance Notes |
---|---|---|---|
DeepSeek Coder | Mini (1.3B), Light (1.5B) | 2-3 GB | 10-15 tokens/sec in FP16 |
TinyLlama | 1.1B | 2 GB | 8-12 tokens/sec in FP16 |
Phi | Phi-1.5 (1.3B), Phi-2 (2.7B) | 1.5-3 GB | Phi-1.5: 8-12 tokens/sec in FP16, Phi-2: 4-8 tokens/sec in FP16 |
Llama 2 | 7B (Quantized to 4-bit) | 3-4 GB | 1-2 tokens/sec in INT4/INT8 |
Mistral | 7B (Quantized to 4-bit) | 3-4 GB | 1-2 tokens/sec in INT4/INT8 |
DeepSeek R1 1.5B Optimizations Recommendations:
- Supports INT4-8 quantization for inference
- Typical throughput: 8-12 tokens/sec in FP16, 12-18 tokens/sec in INT8
- Recommended batch size: 1-2 for real-time applications