About GPUX¶
Docker-like GPU runtime for ML inference with universal GPU compatibility.
What is GPUX?¶
GPUX is a lightweight, Docker-inspired runtime that makes GPU-accelerated ML inference work everywhere - NVIDIA, AMD, Apple Silicon, Intel, and even CPU-only systems.
Key Features¶
- π Universal GPU Support - Works on any GPU (NVIDIA, AMD, Apple, Intel)
- π³ Docker-like UX - Familiar
build,run,servecommands - β‘ Excellent Performance - ONNX Runtime with optimized execution providers
- π§ Simple Configuration - Single
gpux.ymlfile - π¦ Zero Vendor Lock-in - Use ONNX models from any framework
Why GPUX?¶
The Problem¶
ML deployment is fragmented: - Different GPUs need different runtimes (CUDA, ROCm, CoreML, DirectML) - Complex setup and configuration - Vendor lock-in with frameworks
The Solution¶
GPUX provides a unified interface:
# Works everywhere - same commands, any GPU
gpux build .
gpux run model-name --input '{"data": [1,2,3]}'
gpux serve model-name --port 8080
Architecture¶
GPUX is a platform layer built on proven technologies:
βββββββββββββββββββββββββββββββββββββββ
β GPUX (Docker-like UX) β
βββββββββββββββββββββββββββββββββββββββ€
β ONNX Runtime (Core) β
βββββββββββββββββββββββββββββββββββββββ€
β Execution Providers (GPU Backends) β
β TensorRTβCUDAβROCmβCoreMLβDirectML β
βββββββββββββββββββββββββββββββββββββββ€
β Hardware (Any GPU or CPU) β
βββββββββββββββββββββββββββββββββββββββ
Philosophy: We focus on UX and tooling, leveraging ONNX Runtime's battle-tested ML execution.
Technology Stack¶
- Runtime: ONNX Runtime (Microsoft)
- Execution Providers: TensorRT, CUDA, ROCm, CoreML, DirectML, OpenVINO, CPU
- CLI: Typer (Python)
- Serving: FastAPI + Uvicorn
- Configuration: YAML + Pydantic
Project Status¶
- β Production Ready: Built on mature ONNX Runtime
- π Active Development: Regular updates and improvements
- π Open Source: MIT License
Performance¶
GPUX delivers excellent performance through optimized execution providers:
| Hardware | Provider | BERT Throughput | vs CPU |
|---|---|---|---|
| RTX 3080 | TensorRT | 2,400 FPS | 48x |
| M2 Pro | CoreML | 450 FPS | 9x |
| RX 6800 XT | ROCm | 600 FPS | 15x |
Get Involved¶
Use GPUX¶
- π Documentation
- π Quick Start
- π‘ Examples
Contribute¶
- π Report Issues
- π¬ Discussions
- π€ Contributing Guide
Stay Updated¶
- β Star on GitHub
- π° Changelog
- πΊοΈ Roadmap
License¶
GPUX is open source under the MIT License.