How to use this section
The fastest path to a working deployment is to follow the reference architecture. It is an opinionated, validated configuration for a minimal, production-ready Poolside deployment, and ships a Terraform starting point that provisions the AWS infrastructure, the Kubernetes prerequisites, and the Helm install in a singleterraform apply.
If you assemble the AWS foundation with your own infrastructure-as-code, the install guide describes the chart’s expectations so you can match them.
- Reference architecture: Recommended starting point.
- Install the platform on Amazon EKS: Chart-level install steps for self-assembled deployments.
- Configure model inference on Amazon EKS: Local GPU inference or external OpenAI-compatible APIs.
Model inference
You can serve Poolside models from GPU-backed inference workloads inside your Amazon EKS cluster. Deploy theinference-stack Helm chart against your own GPU node group and NVIDIA GPU Operator setup, or use the reference architecture, which provisions both for you. You can also connect Poolside to any external OpenAI-compatible model API instead. See Model inference on Amazon EKS for details.