An external provider for Llama Stack allowing for the use of Podman AI Lab for inference.
-
Ensure Podman Desktop is running and AI Lab extension is loaded
-
Run the Podman AI Lab external provider inside of a container via Podman:
podman run -p 8321:8321 ghcr.io/containers/podman-ai-lab-stack:nightlyThis will start a Llama Stack server which will use port 8321 by default. You can test this works by using the Llama Stack Client:
llama-stack-client models list
llama-stack-client models register <id-of-model-loaded-in-podman-ai-lab>
llama-stack-client inference chat-completion --message "tell me a joke" --streamBy default, Podman AI Lab listens on port 10434, and the Podman AI Lab external provider is configured to access this port by default. If you want to provide another address/port, you can pass the PODMAN_AI_LAB_URL environment variable to the provider, for example:
podman run -p 8321:8321 --env PODMAN_AI_LAB_URL=http://host.containers.internal:10435 ghcr.io/containers/podman-ai-lab-stack:nightly