An external provider for Llama Stack allowing for the use of Podman AI Lab for inference.
-
Ensure Podman Desktop is running and AI Lab extension is loaded
-
Run the Podman AI Lab external provider inside of a container via Podman:
podman run -p 8321:8321 ghcr.io/containers/podman-ai-lab-stack:nightly
This will start a Llama Stack server which will use port 8321 by default. You can test this works by using the Llama Stack Client:
llama-stack-client models list
llama-stack-client models register <id-of-model-loaded-in-podman-ai-lab>
llama-stack-client inference chat-completion --message "tell me a joke" --stream
By default, Podman AI Lab listens on port 10434
, and the Podman AI Lab external provider is configured to access this port by default. If you want to provide another address/port, you can pass the PODMAN_AI_LAB_URL
environment variable to the provider, for example:
podman run -p 8321:8321 --env PODMAN_AI_LAB_URL=http://host.containers.internal:10435 ghcr.io/containers/podman-ai-lab-stack:nightly