HAMi (Heterogeneous AI Computing Virtualization Middleware) formerly known as k8s-vGPU-scheduler, is an 'all-in-one' chart designed to manage Heterogeneous AI Computing Devices in a k8s cluster. It can provide the ability to share Heterogeneous AI devices and provide resource isolation among tasks.
HAMi is committed to improving the utilization rate of heterogeneous computing devices in Kubernetes clusters and providing a unified multiplexing interface for different types of heterogeneous devices.
Zero change upgrade: compatible with default behaviour from Kubernetes.
Jointly initiated by Internet, finance, manufacturing, cloud providers, etc. Target for open governance with CNCF
Integration with mainstream cloud providers; Not tied to proprietary vendor orchestration
Provides hard isolation of resources within containers, task in containers can't use resources that exceed their quota
Provides device-sharing on GPU, MLU, NPU from a variety of manufacturers
Unified monitoring system, Configurable scheduling policies(binpack,spread,etc...)