From 5169f19bbb9513efe1c642e7d54756f8ceade0c1 Mon Sep 17 00:00:00 2001 From: Kyle Mistele Date: Fri, 7 Feb 2025 10:32:04 -0600 Subject: [PATCH] fix: remove network --- docker-compose.ollama.yml | 4 ---- docker-compose.single-gpu.yml | 4 ---- docker-compose.vllm.yml | 4 ---- 3 files changed, 12 deletions(-) diff --git a/docker-compose.ollama.yml b/docker-compose.ollama.yml index 8e2c1701..54306228 100644 --- a/docker-compose.ollama.yml +++ b/docker-compose.ollama.yml @@ -32,7 +32,3 @@ services: - ./node/inference/litellm/litellm_config.ollama.yml:/app/config.yaml ports: - '4000:4000' # TODO do not expose in production - -networks: - default: - name: node-network \ No newline at end of file diff --git a/docker-compose.single-gpu.yml b/docker-compose.single-gpu.yml index e647b804..2646d87a 100644 --- a/docker-compose.single-gpu.yml +++ b/docker-compose.single-gpu.yml @@ -48,7 +48,3 @@ services: - ./node/inference/litellm/litellm_config.single_gpu.yml:/app/config.yaml # this should be the litellm config file ports: - '4000:4000' - -networks: - default: - name: node-network \ No newline at end of file diff --git a/docker-compose.vllm.yml b/docker-compose.vllm.yml index 0968fc57..f22f3190 100644 --- a/docker-compose.vllm.yml +++ b/docker-compose.vllm.yml @@ -250,7 +250,3 @@ services: - ./node/inference/litellm/litellm_config.vllm.yml:/app/config.yaml # this should be the litellm config file ports: - '4000:4000' - -networks: - default: - name: node-network \ No newline at end of file