-
Notifications
You must be signed in to change notification settings - Fork 101
Expand file tree
/
Copy pathweb_demo_vllm.py
More file actions
148 lines (122 loc) · 5.9 KB
/
web_demo_vllm.py
File metadata and controls
148 lines (122 loc) · 5.9 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
import tempfile
import traceback
from pathlib import Path
import gradio as gr
def save_tmp_audio(audio_bytes, cache_dir):
with tempfile.NamedTemporaryFile(dir=cache_dir, delete=False, suffix=".wav") as temp_audio:
temp_audio.write(audio_bytes)
return temp_audio.name
def add_message(chatbot, history, mic, text):
if not mic and not text:
return chatbot, history, "Input is empty"
if text:
chatbot.append({"role": "user", "content": text})
history.append({"role": "human", "content": text})
elif mic and Path(mic).exists():
chatbot.append({"role": "user", "content": {"path": mic}})
history.append({"role": "human", "content": [{"type": "audio", "audio": mic}]})
return chatbot, history, None
def reset_state(system_prompt):
return [], [{"role": "system", "content": system_prompt}]
def predict(chatbot, history, audio_model, token2wav, prompt_wav, cache_dir):
try:
# Request speech response
history.append({"role": "assistant", "content": "<tts_start>", "eot": False})
response, text, audio = audio_model(
history,
max_tokens=2048,
temperature=0.7,
top_p=0.9,
repetition_penalty=1.05,
)
# Convert audio tokens to waveform and append to chat
if audio:
audio = [x for x in audio if x < 6561]
audio_bytes = token2wav(audio, prompt_wav)
audio_path = save_tmp_audio(audio_bytes, cache_dir)
chatbot.append({"role": "assistant", "content": {"path": audio_path}})
history[-1] = {"role": "assistant", "tts_content": response["tts_content"]}
else:
chatbot.append({"role": "assistant", "content": text})
history[-1] = {"role": "assistant", "content": text}
except Exception:
print(traceback.format_exc())
gr.Warning("Some error happened, please try again.")
return chatbot, history
def _launch_demo(args, audio_model, token2wav):
with gr.Blocks(delete_cache=(86400, 86400)) as demo:
gr.Markdown("""<center><font size=8>Step Audio 2 vLLM Demo (Text-only Output)</center>""")
with gr.Row():
system_prompt = gr.Textbox(
label="System Prompt",
value="你的名字叫做小跃,是由阶跃星辰公司训练出来的语音大模型。\n你情感细腻,观察能力强,擅长分析用户的内容,并作出善解人意的回复,说话的过程中时刻注意用户的感受,富有同理心,提供多样的情绪价值。\n今天是2025年8月29日,星期五\n请用默认女声与用户交流。",
lines=2,
)
chatbot = gr.Chatbot(
elem_id="chatbot",
min_height=800,
type="messages",
)
history = gr.State([{"role": "system", "content": system_prompt.value}])
mic = gr.Audio(type="filepath")
text = gr.Textbox(placeholder="Enter message ...")
with gr.Row():
clean_btn = gr.Button("🧹 Clear History (清除历史)")
regen_btn = gr.Button("🤔️ Regenerate (重试)")
submit_btn = gr.Button("🚀 Submit")
def on_submit(chatbot, history, mic, text):
chatbot, history, error = add_message(chatbot, history, mic, text)
if error:
gr.Warning(error)
return chatbot, history, None, None
else:
chatbot, history = predict(chatbot, history, audio_model, token2wav, args.prompt_wav, args.cache_dir)
return chatbot, history, None, None
submit_btn.click(
fn=on_submit,
inputs=[chatbot, history, mic, text],
outputs=[chatbot, history, mic, text],
concurrency_limit=4,
concurrency_id="gpu_queue",
)
clean_btn.click(
fn=reset_state,
inputs=[system_prompt],
outputs=[chatbot, history],
)
def regenerate(chatbot, history):
while chatbot and chatbot[-1]["role"] == "assistant":
chatbot.pop()
while history and history[-1]["role"] == "assistant":
history.pop()
return predict(chatbot, history, audio_model, token2wav, args.prompt_wav, args.cache_dir)
regen_btn.click(
regenerate,
[chatbot, history],
[chatbot, history],
concurrency_id="gpu_queue",
)
demo.queue().launch(
server_port=args.server_port,
server_name=args.server_name,
)
if __name__ == "__main__":
import os
from argparse import ArgumentParser
from stepaudio2vllm import StepAudio2
from token2wav import Token2wav
parser = ArgumentParser()
parser.add_argument("--api-url", type=str, default="http://localhost:8000/v1/chat/completions", help="vLLM OpenAI-compatible endpoint")
parser.add_argument("--model-name", type=str, default="step-audio-2-mini", help="Model name for vLLM serving")
parser.add_argument("--token2wav-path", type=str, default=None, help="Path to token2wav directory (defaults to Step-Audio-2-mini/token2wav)")
parser.add_argument("--prompt-wav", type=str, default="assets/default_female.wav", help="Prompt wave for the assistant.")
parser.add_argument("--cache-dir", type=str, default="/tmp/stepaudio2", help="Cache directory for generated audio.")
parser.add_argument("--server-port", type=int, default=7862, help="Demo server port.")
parser.add_argument("--server-name", type=str, default="0.0.0.0", help="Demo server name.")
args = parser.parse_args()
os.environ["GRADIO_TEMP_DIR"] = args.cache_dir
Path(args.cache_dir).mkdir(parents=True, exist_ok=True)
audio_model = StepAudio2(args.api_url, args.model_name)
token2wav_path = args.token2wav_path or "Step-Audio-2-mini/token2wav"
token2wav = Token2wav(token2wav_path)
_launch_demo(args, audio_model, token2wav)