-
services:
tabby:
restart: always
image: tabbyml/tabby
entrypoint: /opt/tabby/bin/tabby-cpu
command: serve --model StarCoder-1B --chat-model Qwen2-1.5B-Instruct
volumes:
- ".data/tabby:/data"
ports:
- 8080:8080 which is document here https://tabby.tabbyml.com/docs/quick-start/installation/docker-compose/ wont work
Originally posted by @b-reich in #2082 (comment) |
Beta Was this translation helpful? Give feedback.
Replies: 19 comments
-
Hi, thanks for reporting the issue. For a workaround, I recommend use the linux binary distribution directly: https://tabby.tabbyml.com/docs/quick-start/installation/linux/#download-the-release |
Beta Was this translation helpful? Give feedback.
-
I also encountered the same error😭 |
Beta Was this translation helpful? Give feedback.
-
The issue seems to be related to the llama-server, the |
Beta Was this translation helpful? Give feedback.
-
Submitted the pull request #2711 . In the meanwhile you can use my temporary image If you're using docker compose you can use the following snippet: version: '3.8'
services:
tabby:
restart: always
image: tabbyml/tabby
command: serve --model StarCoder-1B --chat-model Qwen2-1.5B-Instruct --device cuda
volumes:
- "$HOME/.tabby:/data"
ports:
- 8080:8080
environment:
- PATH=/usr/local/cuda/bin:$PATH
- LD_LIBRARY_PATH=/usr/local/cuda/lib64:/usr/local/cuda/compat:$LD_LIBRARY_PATH
- NVIDIA_VISIBLE_DEVICES=all
- NVIDIA_DRIVER_CAPABILITIES=compute,utility
runtime: nvidia
deploy:
resources:
reservations:
devices:
- driver: nvidia
count: 1
capabilities: [gpu] If you're using docker you can use the following snippet: docker run -it --gpus all \
-p 8080:8080 \
-v $HOME/.tabby:/data \
-e PATH=/usr/local/cuda/bin:$PATH \
-e LD_LIBRARY_PATH=/usr/local/cuda/lib64:/usr/local/cuda/compat:$LD_LIBRARY_PATH \
tabbyml/tabby serve --model StarCoder-1B --chat-model Qwen2-1.5B-Instruct --device cuda |
Beta Was this translation helpful? Give feedback.
-
@0x4139 nope this is not my issue. I want to run it without a GPU. Just CPU Mode. |
Beta Was this translation helpful? Give feedback.
-
Same error here. |
Beta Was this translation helpful? Give feedback.
-
For those experiencing the issue, please refer to the comment at #2634 (comment) to see if it resolves the problem for you. If it doesn't, feel free to share your experiences. Thank you! |
Beta Was this translation helpful? Give feedback.
-
The issues are related, the binary won't start even in cpu mode due to the fact that the cuda libraries are not linked. Just tested it now, and it works also in cpu mode. |
Beta Was this translation helpful? Give feedback.
-
@0x4139 ur docker command and compose uses different images. |
Beta Was this translation helpful? Give feedback.
-
That is the point, i mentioned that i created a temporary image with the LD path fix, that works both on CPU and GPU. If the image works for you as well, probably @wsxiaoys will merge the fix. |
Beta Was this translation helpful? Give feedback.
-
I'm experiencing similar issue, but for me Docker image works fine, it's Linux release that doesn't work. Error:
I'm using command: Adding env as suggested #2634 (comment) doesn't help: EndeavourOS, tabby P.S. Is it fine to discuss it here or should I open new issue? |
Beta Was this translation helpful? Give feedback.
-
Be sure you have installed |
Beta Was this translation helpful? Give feedback.
-
Thank that fixed it. Now I'm getting:
Seems similar to this #2803 |
Beta Was this translation helpful? Give feedback.
-
Could you provide a broader view of the logs, as well as your tabby configuration? |
Beta Was this translation helpful? Give feedback.
-
I'm on EndeavourOS and I've downloaded https://github.com/TabbyML/tabby/releases/tag/v0.14.0 / Command
It just goes on forever. Here's
Where can I find tabby configuration so I can provide it? I looked at |
Beta Was this translation helpful? Give feedback.
-
Seems to be related to some flags passed to the |
Beta Was this translation helpful? Give feedback.
-
tabby build the image with cuda by default, Line 42 in 48d9c08 that's why llama-cpp-server looks for the libcuda, and failed to start if no GPU existed. the libcuda is mounted at runtime by nvidia-container-runtime. maybe we need a cpu Dockerfile to build the CPU image? it can also largely reduce the image size without the cuda dependencies. WDYT @wsxiaoys |
Beta Was this translation helpful? Give feedback.
-
Thank you that fixed the problem. v15 was released but I get the same error on it. Is there any issue/pr that related to it? So I can monitor when it's safe to upgrade. |
Beta Was this translation helpful? Give feedback.
-
There is a merge request here #2711 |
Beta Was this translation helpful? Give feedback.
Hi, thanks for reporting the issue. For a workaround, I recommend use the linux binary distribution directly: https://tabby.tabbyml.com/docs/quick-start/installation/linux/#download-the-release