Skip to content

Commit c3dad34

Browse files
committed
lint
Signed-off-by: Ekagra Ranjan <3116519+ekagra-ranjan@users.noreply.github.com>
1 parent ea38d5a commit c3dad34

File tree

5 files changed

+24
-28
lines changed

5 files changed

+24
-28
lines changed

examples/online_serving/stable_audio/README.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -188,11 +188,11 @@ python stable_audio_client.py \
188188
## Tips
189189

190190
1. **Audio Length**: Keep under 47 seconds for `stable-audio-open-1.0`
191-
2. **Quality vs Speed**:
191+
2. **Quality vs Speed**:
192192
- 50 steps: Fast, decent quality
193193
- 100 steps: Good balance (default)
194194
- 150+ steps: High quality, slower
195-
3. **Guidance Scale**:
195+
3. **Guidance Scale**:
196196
- Lower (3-5): More creative/varied
197197
- Default (7): Good balance
198198
- Higher (10+): More literal to prompt

examples/online_serving/stable_audio/stable_audio_client.py

Lines changed: 9 additions & 11 deletions
Original file line numberDiff line numberDiff line change
@@ -29,9 +29,7 @@
2929

3030

3131
def parse_args():
32-
parser = argparse.ArgumentParser(
33-
description="Generate audio with Stable Audio via OpenAI-compatible API"
34-
)
32+
parser = argparse.ArgumentParser(description="Generate audio with Stable Audio via OpenAI-compatible API")
3533
parser.add_argument(
3634
"--api_url",
3735
default="http://localhost:8000/v1/audio/speech",
@@ -93,15 +91,15 @@ def parse_args():
9391

9492
def generate_audio(args):
9593
"""Generate audio using the API."""
96-
94+
9795
# Build request payload
9896
payload = {
9997
"input": args.text,
10098
"audio_length": args.audio_length,
10199
"audio_start": args.audio_start,
102100
"response_format": args.response_format,
103101
}
104-
102+
105103
# Add optional parameters
106104
if args.negative_prompt:
107105
payload["negative_prompt"] = args.negative_prompt
@@ -111,7 +109,7 @@ def generate_audio(args):
111109
payload["num_inference_steps"] = args.num_inference_steps
112110
if args.seed is not None:
113111
payload["seed"] = args.seed
114-
112+
115113
print(f"\n{'=' * 60}")
116114
print("Stable Audio - Text-to-Audio Generation")
117115
print(f"{'=' * 60}")
@@ -125,7 +123,7 @@ def generate_audio(args):
125123
print(f"Seed: {args.seed}")
126124
print(f"Output: {args.output}")
127125
print(f"{'=' * 60}\n")
128-
126+
129127
try:
130128
# Make the API request
131129
print("Generating audio...")
@@ -135,21 +133,21 @@ def generate_audio(args):
135133
headers={"Content-Type": "application/json"},
136134
timeout=300, # 5 minute timeout for long generations
137135
)
138-
136+
139137
# Check for errors
140138
if response.status_code != 200:
141139
print(f"Error: API returned status code {response.status_code}")
142140
print(f"Response: {response.text}")
143141
return False
144-
142+
145143
# Save the audio
146144
with open(args.output, "wb") as f:
147145
f.write(response.content)
148-
146+
149147
print(f"✓ Audio saved to {args.output}")
150148
print(f" File size: {len(response.content) / 1024:.1f} KB")
151149
return True
152-
150+
153151
except requests.exceptions.Timeout:
154152
print("Error: Request timed out. Try reducing inference steps or audio length.")
155153
return False

vllm_omni/entrypoints/openai/api_server.py

Lines changed: 5 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -139,7 +139,8 @@ class _DiffusionServingModels:
139139
"""
140140

141141
class _NullModelConfig:
142-
def __getattr__(self, name): return None
142+
def __getattr__(self, name):
143+
return None
143144

144145
class _Unsupported:
145146
def __init__(self, name: str):
@@ -151,7 +152,6 @@ def __call__(self, *args, **kwargs):
151152
def __getattr__(self, attr):
152153
raise NotImplementedError(f"{self.name}.{attr} is not supported in diffusion mode")
153154

154-
155155
def __init__(self, base_model_paths: list[BaseModelPath]) -> None:
156156
self._base_model_paths = base_model_paths
157157
self.model_config = self._NullModelConfig()
@@ -429,8 +429,8 @@ async def omni_init_app_state(
429429
)
430430

431431
state.openai_serving_speech = OmniOpenAIServingSpeech.for_diffusion(
432-
engine_client,
433-
state.openai_serving_models,
432+
engine_client,
433+
state.openai_serving_models,
434434
request_logger=request_logger,
435435
model_name=model_name,
436436
)
@@ -710,10 +710,7 @@ async def omni_init_app_state(
710710
)
711711

712712
state.openai_serving_speech = OmniOpenAIServingSpeech(
713-
engine_client,
714-
state.openai_serving_models,
715-
request_logger=request_logger,
716-
model_name=model_name
713+
engine_client, state.openai_serving_models, request_logger=request_logger, model_name=model_name
717714
)
718715

719716
state.enable_server_load_tracking = args.enable_server_load_tracking

vllm_omni/entrypoints/openai/audio_utils_mixin.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -44,7 +44,7 @@ def create_audio(self, audio_obj: CreateAudio) -> AudioResponse:
4444
f"Unsupported audio tensor dimension: {audio_tensor.ndim}. "
4545
"Only mono (1D) and stereo (2D) are supported."
4646
)
47-
47+
4848
if audio_tensor.ndim == 2 and audio_tensor.shape[0] == 2:
4949
# Convert from [channels, samples] to [samples, channels]
5050
audio_tensor = audio_tensor.T

vllm_omni/entrypoints/openai/serving_speech.py

Lines changed: 7 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -1,9 +1,9 @@
11
import asyncio
22
from typing import Any
33

4+
import torch
45
from fastapi import Request
56
from fastapi.responses import Response
6-
import torch
77
from vllm.entrypoints.openai.engine.serving import OpenAIServing
88
from vllm.logger import init_logger
99
from vllm.utils import random_uuid
@@ -247,7 +247,7 @@ async def create_speech(
247247
# Handle Stable Audio models
248248
# Stable Audio uses diffusion, needs different parameters
249249
default_sr = 44100 # Default sample rate for Stable Audio
250-
250+
251251
# Build prompt for Stable Audio
252252
prompt = {
253253
"prompt": request.input,
@@ -257,19 +257,20 @@ async def create_speech(
257257

258258
# Build sampling params for diffusion
259259
sampling_params_list[0].num_outputs_per_prompt = 1
260-
260+
261261
# Create generator if seed provided
262262
if request.seed is not None:
263263
from vllm_omni.platforms import current_omni_platform
264+
264265
generator = torch.Generator(device=current_omni_platform.device_type).manual_seed(request.seed)
265266
sampling_params_list[0].generator = generator
266-
267+
267268
if request.guidance_scale is not None:
268269
sampling_params_list[0].guidance_scale = request.guidance_scale
269-
270+
270271
if request.num_inference_steps is not None:
271272
sampling_params_list[0].num_inference_steps = request.num_inference_steps
272-
273+
273274
# Set up audio duration parameters
274275
if request.audio_length is not None:
275276
audio_length = request.audio_length

0 commit comments

Comments
 (0)