Welcome to exalsius
exalsius is a platform for building and managing GPU infrastructure for AI workloads. It lets you assemble clusters from any mix of cloud providers and on-premise machines, then run development environments, inference endpoints, and distributed training jobs on top of them.
What you can do with exalsius
- Bring your own infrastructure: Add cloud instances or bare-metal machines to a shared node pool with node management.
- Run Kubernetes anywhere: Deploy multi-cloud and hybrid clusters for AI development and serving with cluster deployments.
- Launch AI workloads fast: Start notebooks, dev pods, inference endpoints, and training jobs with workspaces.
- Observe everything in one place: View metrics, logs, and traces with built-in observability dashboards.
How it works
exalsius follows a simple three-step workflow:
- Nodes: Connect the infrastructure you want to use.
- Clusters: Provision Kubernetes on selected nodes.
- Workspaces: Launch the environments and services your team needs.
Each step builds on the previous one. The quickstart walks you through the full flow.