Skip to main content
Version: v4.7 Stable

Run:ai

Run:ai is a GPU orchestration platform that schedules and manages AI and machine learning workloads across Kubernetes clusters. vCluster is a certified Kubernetes Distribution for hosting Run:ai that lets platform teams share GPU infrastructure across isolated tenants without provisioning separate physical clusters for each team.

Compatibility has been verified with the following versions.

ComponentVersion
Kubernetesv1.34
vClusterv0.31
Run:aiv2.24

Deployment models​

vCluster supports all three deployment models with Run:ai:

ModelDescriptionUse case
Shared nodesTenants share host cluster nodes with label-based scheduling. Each tenant gets a separate vCluster with its own Kubernetes API.Trusted tenants, cost-efficient GPU sharing
Private nodesEach tenant gets a dedicated vCluster with auto-provisioned private nodes.Untrusted tenants, strict compliance requirements
StandaloneA single-tenant deployment where Run:ai manages the full node pool within one vCluster. No multi-tenancy overhead.Single team, dedicated GPU pool

Installation​

The Run:ai certified stack provisions a vCluster together with the NVIDIA GPU Operator and Run:ai in a single, tested deployment. See the certified-stacks repository for prerequisites, configuration options, and step-by-step setup instructions.


Run:ai certified stack