File tree Expand file tree Collapse file tree 3 files changed +50
-4
lines changed
components/app-configs/gpu-verification-amd/base Expand file tree Collapse file tree 3 files changed +50
-4
lines changed Original file line number Diff line number Diff line change 2222 - -c
2323 - |
2424 #!/bin/bash
25-
26- rocm -smi
25+ amd-smi list
26+ amd -smi
2727 while true
2828 do
2929 python -c "import torch; print('ROCm Available:', torch.cuda.is_available()); print('GPU Count:', torch.cuda.device_count())"
Original file line number Diff line number Diff line change 1+ apiVersion : v1
2+ kind : Pod
3+ metadata :
4+ labels :
5+ app : amd-rocm-vllm
6+ name : amd-rocm-vllm
7+ spec :
8+ restartPolicy : OnFailure
9+ containers :
10+ - name : minion
11+ image : docker.io/rocm/vllm:rocm7.0.0_vllm_0.11.1_20251103
12+ command :
13+ - /bin/bash
14+ - -c
15+ - |
16+ #!/bin/bash
17+
18+ model=Qwen/Qwen3-8B
19+ tp=1
20+ dtype=auto
21+ kv_cache_dtype=auto
22+ max_num_seqs=1024
23+ max_num_batched_tokens=40960
24+ max_model_len=8192
25+
26+ vllm serve $model \
27+ -tp $tp \
28+ --dtype $dtype \
29+ --kv-cache-dtype $kv_cache_dtype \
30+ --max-num-seqs $max_num_seqs \
31+ --max-num-batched-tokens $max_num_batched_tokens \
32+ --max-model-len $max_model_len \
33+ --no-enable-prefix-caching \
34+ --swap-space 16 \
35+ --disable-log-requests
36+ volumeMounts :
37+ - name : shared
38+ mountPath : /dev/shm
39+ resources :
40+ limits :
41+ amd.com/gpu : 1
42+ volumes :
43+ - emptyDir :
44+ medium : Memory
45+ sizeLimit : 1Gi
46+ name : shared
Original file line number Diff line number Diff line change 1414 - -c
1515 - |
1616 #!/bin/bash
17-
18- rocm -smi
17+ amd-smi list
18+ amd -smi
1919 while true
2020 do
2121 python -c "import torch; print('ROCm Available:', torch.cuda.is_available()); print('GPU Count:', torch.cuda.device_count())"
You can’t perform that action at this time.
0 commit comments