|
| 1 | +#!/usr/bin/env bash |
| 2 | +# Shared deploy path for llm-d-infra nightly reusables (CKS + OpenShift). |
| 3 | +# Invoked via: make nightly-deploy-wva-guide (sets LLM_D_NIGHTLY_PLATFORM=cks|openshift). |
| 4 | +set -euo pipefail |
| 5 | + |
| 6 | +ROOT="${1:-.}" |
| 7 | +cd "$ROOT" |
| 8 | + |
| 9 | +PLATFORM="${LLM_D_NIGHTLY_PLATFORM:?LLM_D_NIGHTLY_PLATFORM must be cks or openshift}" |
| 10 | +if [[ "$PLATFORM" != cks && "$PLATFORM" != openshift ]]; then |
| 11 | + echo "LLM_D_NIGHTLY_PLATFORM must be cks or openshift (got: $PLATFORM)" >&2 |
| 12 | + exit 1 |
| 13 | +fi |
| 14 | + |
| 15 | +if [[ -n "${GITHUB_WORKSPACE:-}" && ! -d llm-d && -d "$GITHUB_WORKSPACE/guides" ]]; then |
| 16 | + ln -sfn "$GITHUB_WORKSPACE" llm-d |
| 17 | + echo "Symlinked $ROOT/llm-d -> $GITHUB_WORKSPACE" |
| 18 | +fi |
| 19 | + |
| 20 | +if [[ "$PLATFORM" == cks ]]; then |
| 21 | + for f in deploy/lib/deploy_prometheus_kube_stack.sh deploy/kubernetes/install.sh; do |
| 22 | + if [[ -f "$f" ]] && grep -q 'helm upgrade --install kube-prometheus-stack' "$f" && ! grep -q 'nodeExporter.enabled=false' "$f"; then |
| 23 | + perl -pi -e 's/helm upgrade --install kube-prometheus-stack/helm upgrade --install kube-prometheus-stack --set nodeExporter.enabled=false/g' "$f" |
| 24 | + echo "Patched $f: nodeExporter.enabled=false (CKS nightly)" |
| 25 | + fi |
| 26 | + done |
| 27 | +fi |
| 28 | + |
| 29 | +export INSTALL_GATEWAY_CTRLPLANE="${INSTALL_GATEWAY_CTRLPLANE:-false}" |
| 30 | +export BENCHMARK_MODE="${BENCHMARK_MODE:-false}" |
| 31 | +export E2E_TESTS_ENABLED="${E2E_TESTS_ENABLED:-true}" |
| 32 | +export NAMESPACE_SCOPED="${NAMESPACE_SCOPED:-false}" |
| 33 | +export DEPLOY_WVA="${DEPLOY_WVA:-true}" |
| 34 | +export DEPLOY_PROMETHEUS="${DEPLOY_PROMETHEUS:-true}" |
| 35 | +export DEPLOY_PROMETHEUS_ADAPTER="${DEPLOY_PROMETHEUS_ADAPTER:-true}" |
| 36 | +export DEPLOY_VA="${DEPLOY_VA:-false}" |
| 37 | +export DEPLOY_HPA="${DEPLOY_HPA:-false}" |
| 38 | +export SCALER_BACKEND="${SCALER_BACKEND:-keda}" |
| 39 | +export ENABLE_SCALE_TO_ZERO="${ENABLE_SCALE_TO_ZERO:-true}" |
| 40 | +export POOL_GROUP="${POOL_GROUP:-inference.networking.k8s.io}" |
| 41 | + |
| 42 | +if [[ "$PLATFORM" == openshift ]]; then |
| 43 | + export MONITORING_NAMESPACE="${MONITORING_NAMESPACE:-openshift-user-workload-monitoring}" |
| 44 | + export WVA_METRICS_SECURE="${WVA_METRICS_SECURE:-false}" |
| 45 | + export ENVIRONMENT=openshift |
| 46 | + ./deploy/install.sh \ |
| 47 | + --model "${MODEL_ID:-unsloth/Meta-Llama-3.1-8B}" \ |
| 48 | + --accelerator "${ACCELERATOR_TYPE:-H100}" \ |
| 49 | + --release-name "${WVA_RELEASE_NAME:-workload-variant-autoscaler}" \ |
| 50 | + --environment openshift |
| 51 | +else |
| 52 | + export ENVIRONMENT=kubernetes |
| 53 | + ./deploy/install.sh \ |
| 54 | + --model "${MODEL_ID:-unsloth/Meta-Llama-3.1-8B}" \ |
| 55 | + --accelerator "${ACCELERATOR_TYPE:-H100}" \ |
| 56 | + --release-name "${WVA_RELEASE_NAME:-workload-variant-autoscaler}" |
| 57 | +fi |
0 commit comments