This guide covers installing Workload-Variant-Autoscaler (WVA) on your Kubernetes cluster.
- Kubernetes v1.32.0 or later with administrator access or namespace-level permissions
- Helm 3.x
- kubectl configured to access your cluster
Using kustomize for more control:
# Install CRDs
make install
# Deploy the controller
make deploy IMG=quay.io/llm-d/llm-d-workload-variant-autoscaler:latestSee the Helm Installation for detailed instructions.
Verify the installation:
kubectl get pods -n workload-variant-autoscaler-systemSee the comprehensive deployment guide for detailed instructions.
WVA can work with existing autoscalers:
For HPA integration: See HPA Integration Guide
For KEDA integration: See KEDA Integration Guide
-
Check controller is running:
kubectl get deployment -n workload-variant-autoscaler-system
-
Verify CRDs are installed:
kubectl get crd variantautoscalings.llmd.ai
-
Check controller logs:
kubectl logs -n workload-variant-autoscaler-system \ deployment/workload-variant-autoscaler-controller-manager
Kustomize:
make undeploy
make uninstall # Remove CRDsHelm:
helm uninstall workload-variant-autoscaler -n workload-variant-autoscaler-systemController not starting:
- Check if CRDs are installed:
kubectl get crd - Verify RBAC permissions
- Check controller logs for errors
Metrics not appearing:
- Ensure Prometheus ServiceMonitor is created
- Verify Prometheus has proper RBAC to scrape metrics
- Check network policies aren't blocking metrics endpoint
See Also: