diff --git a/apps/ollama-amd/config.json b/apps/ollama-amd/config.json index f536675336..255ed57a67 100644 --- a/apps/ollama-amd/config.json +++ b/apps/ollama-amd/config.json @@ -5,8 +5,8 @@ "exposable": true, "port": 11434, "id": "ollama-amd", - "tipi_version": 48, - "version": "0.5.3-rocm", + "tipi_version": 49, + "version": "0.5.4-rocm", "categories": ["ai"], "description": "Get up and running with Llama 3, Mistral, Gemma, and other large language models.", "short_desc": "LLMs inference server with OpenAI compatible API", @@ -16,5 +16,5 @@ "form_fields": [], "supported_architectures": ["arm64", "amd64"], "created_at": 1691943801422, - "updated_at": 1734450622000 + "updated_at": 1734465908000 } diff --git a/apps/ollama-amd/docker-compose.yml b/apps/ollama-amd/docker-compose.yml index 23998b24a9..bd6c81e9c5 100755 --- a/apps/ollama-amd/docker-compose.yml +++ b/apps/ollama-amd/docker-compose.yml @@ -2,7 +2,7 @@ version: '3.7' services: ollama-amd: - image: ollama/ollama:0.5.3-rocm + image: ollama/ollama:0.5.4-rocm restart: unless-stopped container_name: ollama-amd environment: