Chatbox AI app backend support #1047
magikRUKKOLA
started this conversation in
Show and tell
Replies: 1 comment 6 replies
-
Is there a way to fold the <think> produced by R1 series? My local deployment is too thinkable that its <think> is longer than its answer... |
Beta Was this translation helpful? Give feedback.
6 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
In case someone would want to use ktransformers (multiple ones, running at different machines) as a backend for apps like Chatbox AI that support Ollama API they could use the following.
[EDIT] (link to the latest version of nginx config updated)
chatboxai/chatbox#2221 (comment)
the picture below shows the usage of the llms from three different backends (ollama and 2 ktransformers).
Beta Was this translation helpful? Give feedback.
All reactions