workload-variant-autoscaler/workload-variant-autoscaler 0.5.0 Public Latest
Helm chart for Workload-Variant-Autoscaler (WVA) - GPU-aware autoscaler for LLM inference workloads
Install from the command line
Learn more about packages
$ docker pull ghcr.io/llm-d-incubation/workload-variant-autoscaler/workload-variant-autoscaler:0.5.0
Recent tagged image versions
- 189 Version downloads
- 8 Version downloads
- 138 Version downloads
- 58 Version downloads
- 91 Version downloads
Loading
Sorry, something went wrong.
Last published
2 months ago
Total downloads