Replies: 2 comments 1 reply
-
|
@sujee please feel free to tag anyone else in that you think could help with this discussion |
Beta Was this translation helpful? Give feedback.
-
|
Hi Steve, I am also working on adding vLLM support as a contribution and it seems non-trivial to support (especially for CPU-only machines). Were you able to add vLLM to Allycat? Also, is there a specific reason for using vLLM instead of Ollama (perhaps you want local GPU inference)? Ollama seems to be pretty good for inference on local CPU-only machines (correct me if I am wrong). I have looked into adding vLLM and have noted my thoughts here. Any help is appreciated. |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
Hey folks - I just spun up the default Allycat with Docling and several different Open Weights models. Kudos to those who have contributed before - basically worked right out of the box and nice and simple flow. Also love that you also included a pure code version for every jupyter notebook.
Couple of questions:
a. Is that welcome or should I just leave it in my own fork?
b. If it is welcome, how should I approach contributing it back?
Beta Was this translation helpful? Give feedback.
All reactions