From b1f39fadb290b025c1eebd5213a38bcaf82bf398 Mon Sep 17 00:00:00 2001 From: Houssem Dellai Date: Fri, 20 Sep 2024 09:18:21 +0200 Subject: [PATCH] 510 --- 510_ai_ollama_k8s/Readme.md | 2 ++ 1 file changed, 2 insertions(+) diff --git a/510_ai_ollama_k8s/Readme.md b/510_ai_ollama_k8s/Readme.md index 175fc05..528f2ea 100644 --- a/510_ai_ollama_k8s/Readme.md +++ b/510_ai_ollama_k8s/Readme.md @@ -94,6 +94,8 @@ Here are some example models that can be used in `ollama` [available here](https ## Important notes - The `ollama` server is running only on CPU. However, it can also run on GPU or also NPU. +- As LLM models size are large, it is recommended to use a VM with large disk space. +- During the inference, the model will consume a lot of memory and CPU. It is recommended to use a VM with a large memory and CPU. ## References