Skip to content
Discussion options

You must be logged in to vote

Hi @valerio-bozzolan ,
huggingface "endpoints" are a service people use for production (dedicated paid endpoints)

We had an adapter for the public API but for some reason (I remember low availability and too much variance in model inputs/outputs) we ditched it.

Not sure at the moment what works with HF, and what not, and I'm not investing time in it. Most people running local models use Ollama or vLLM, or many other tools you can use via OpenAI-compatible adapter.

Still you can write your own LLM adapter (see plugins already published about Groq or TogetherAI)

Peace and thank you for playing with the cat ;)
Welcome

Replies: 1 comment 1 reply

Comment options

You must be logged in to vote
1 reply
@valerio-bozzolan
Comment options

Answer selected by valerio-bozzolan
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
2 participants