We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
1 parent 7d2124b commit ebe45d9Copy full SHA for ebe45d9
test/vllm/ec2/utils/test_agents.py
@@ -23,24 +23,16 @@ class AnalysisResult(BaseModel):
23
24
25
def test_direct_completion():
26
- """Test direct API calls to VLLM"""
27
client = OpenAI(
28
api_key=OPENAI_API_KEY,
29
base_url=OPENAI_API_BASE,
30
)
31
+ prompt = "What are the main benefits of using VLLM for inference?"
32
+
33
chat_response = client.completions.create(
34
model=MODEL_NAME,
- prompt=[
35
- {
36
- "role": "system",
37
- "content": "You are a helpful assistant.",
38
- },
39
40
- "role": "user",
41
- "content": "What are the main benefits of using VLLM for inference?",
42
43
- ],
+ prompt=prompt,
44
temperature=0.7,
45
max_tokens=512,
46
0 commit comments