-
Notifications
You must be signed in to change notification settings - Fork 136
Open
Labels
bugSomething isn't workingSomething isn't working
Description
Description
Docker: nvcr.io/nvidia/tritonserver:23.04-py3
Gpu: A100
How can i stop bi-direction streaming(decoupled mode)?
- I want to stop model inference(streaming response) when the user disconnects or according to certain conditions, but I don't know how to do that at the moment.
Reference
- https://github.com/triton-inference-server/server/issues/4344
- https://github.com/triton-inference-server/server/issues/5833#issuecomment-1561318646Reproduced Steps
-Metadata
Metadata
Assignees
Labels
bugSomething isn't workingSomething isn't working