Skip to content

Commit 980dee7

Browse files
committed
update: gpu verification
1 parent d7d5bb8 commit 980dee7

File tree

3 files changed

+50
-4
lines changed

3 files changed

+50
-4
lines changed

components/app-configs/gpu-verification-amd/base/deployment.yaml

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -22,8 +22,8 @@ spec:
2222
- -c
2323
- |
2424
#!/bin/bash
25-
26-
rocm-smi
25+
amd-smi list
26+
amd-smi
2727
while true
2828
do
2929
python -c "import torch; print('ROCm Available:', torch.cuda.is_available()); print('GPU Count:', torch.cuda.device_count())"
Lines changed: 46 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,46 @@
1+
apiVersion: v1
2+
kind: Pod
3+
metadata:
4+
labels:
5+
app: amd-rocm-vllm
6+
name: amd-rocm-vllm
7+
spec:
8+
restartPolicy: OnFailure
9+
containers:
10+
- name: minion
11+
image: docker.io/rocm/vllm:rocm7.0.0_vllm_0.11.1_20251103
12+
command:
13+
- /bin/bash
14+
- -c
15+
- |
16+
#!/bin/bash
17+
18+
model=Qwen/Qwen3-8B
19+
tp=1
20+
dtype=auto
21+
kv_cache_dtype=auto
22+
max_num_seqs=1024
23+
max_num_batched_tokens=40960
24+
max_model_len=8192
25+
26+
vllm serve $model \
27+
-tp $tp \
28+
--dtype $dtype \
29+
--kv-cache-dtype $kv_cache_dtype \
30+
--max-num-seqs $max_num_seqs \
31+
--max-num-batched-tokens $max_num_batched_tokens \
32+
--max-model-len $max_model_len \
33+
--no-enable-prefix-caching \
34+
--swap-space 16 \
35+
--disable-log-requests
36+
volumeMounts:
37+
- name: shared
38+
mountPath: /dev/shm
39+
resources:
40+
limits:
41+
amd.com/gpu: 1
42+
volumes:
43+
- emptyDir:
44+
medium: Memory
45+
sizeLimit: 1Gi
46+
name: shared

components/app-configs/gpu-verification-amd/base/pod.yaml

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -14,8 +14,8 @@ spec:
1414
- -c
1515
- |
1616
#!/bin/bash
17-
18-
rocm-smi
17+
amd-smi list
18+
amd-smi
1919
while true
2020
do
2121
python -c "import torch; print('ROCm Available:', torch.cuda.is_available()); print('GPU Count:', torch.cuda.device_count())"

0 commit comments

Comments
 (0)