Replies: 4 comments
-
|
You can definitely stream If you use ctx.write_event_to_stream in your workflow, you can access these streamed events with the client |
Beta Was this translation helpful? Give feedback.
-
|
Still working on proper docs, but if you are using the old client
Finding the example for the newer client... |
Beta Was this translation helpful? Give feedback.
-
|
Ah, newer client is here |
Beta Was this translation helpful? Give feedback.
-
|
Streaming outside of llama-deploy is slightly different You still use the same ctx.write_event_to_stream method, but how you get the stream is different |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
In the example provided at Llama Deploy Python Fullstack, the final output of the workflow is non-streaming. It only produces results once all generated tokens are complete.
As a result, I had to create my own FastAPI service.
Question:
Is there a more suitable modification to the workflow or llama-deploy that allows for direct and true asynchronous output without having to write my own FastAPI service?
Objective:
My main goal is to deliver the results of the workflow to users as quickly as possible.
Beta Was this translation helpful? Give feedback.
All reactions