Skip to content

Commit 841464c

Browse files
Update Conceptual_Guide/Part_8-semantic_caching/README.md
Co-authored-by: Kris Hung <[email protected]>
1 parent 210a400 commit 841464c

File tree

1 file changed

+1
-1
lines changed
  • Conceptual_Guide/Part_8-semantic_caching

1 file changed

+1
-1
lines changed

Conceptual_Guide/Part_8-semantic_caching/README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -286,7 +286,7 @@ sys 0m0.015s
286286
Now, let's try a different response, but keep the semantics:
287287

288288
```bash
289-
time curl -X POST localhost:8000/v2/models/vllm_model/generate -d '{"text_input": "How do I set up model repository for Triton Inference Server?", "parameters": {"stream": false, "temperature": 0, "max_tokens":100}, "exclude_input_in_output":true}
289+
time curl -X POST localhost:8000/v2/models/vllm_model/generate -d '{"text_input": "How do I set up model repository for Triton Inference Server?", "parameters": {"stream": false, "temperature": 0, "max_tokens":100}, "exclude_input_in_output":true}'
290290
```
291291

292292
Upon success, you should see a response from the server like this one:

0 commit comments

Comments
 (0)