From 89477ea9d3a83181b0222b732a81c71db9edf142 Mon Sep 17 00:00:00 2001 From: Javier Martinez Date: Mon, 12 Aug 2024 08:23:16 +0200 Subject: [PATCH] fix: naming image and ollama-cpu (#2056) --- docker-compose.yaml | 7 +++---- 1 file changed, 3 insertions(+), 4 deletions(-) diff --git a/docker-compose.yaml b/docker-compose.yaml index be0ee63f10..c2ef0f6d7d 100644 --- a/docker-compose.yaml +++ b/docker-compose.yaml @@ -7,7 +7,7 @@ services: # Private-GPT service for the Ollama CPU and GPU modes # This service builds from an external Dockerfile and runs the Ollama mode. private-gpt-ollama: - image: ${PGPT_IMAGE:-zylonai/private-gpt}${PGPT_TAG:-0.6.2}-ollama # x-release-please-version + image: ${PGPT_IMAGE:-zylonai/private-gpt}:${PGPT_TAG:-0.6.2}-ollama # x-release-please-version build: context: . dockerfile: Dockerfile.ollama @@ -31,7 +31,7 @@ services: # Private-GPT service for the local mode # This service builds from a local Dockerfile and runs the application in local mode. private-gpt-llamacpp-cpu: - image: ${PGPT_IMAGE:-zylonai/private-gpt}${PGPT_TAG:-0.6.2}-llamacpp-cpu # x-release-please-version + image: ${PGPT_IMAGE:-zylonai/private-gpt}:${PGPT_TAG:-0.6.2}-llamacpp-cpu # x-release-please-version build: context: . dockerfile: Dockerfile.llamacpp-cpu @@ -57,7 +57,6 @@ services: ollama: image: traefik:v2.10 ports: - - "11435:11434" - "8081:8080" command: - "--providers.file.filename=/etc/router.yml" @@ -84,7 +83,7 @@ services: - ./models:/root/.ollama profiles: - "" - - ollama + - ollama-cpu # Ollama service for the CUDA mode ollama-cuda: