Skip to content

Why Does bitnet.cpp Throw an Error When I Try to Run Inference With a Model Downloaded From Hugging Face? #183

Discussion options

You must be logged in to vote

This issue usually stems from not specifying the full path to the exact .gguf model file in your inference command. bitnet.cpp expects the -m argument to point directly to the .gguf model (e.g., models/BitNet-b1.58-2B-4T/ggml-model-i2_s.gguf), not just the model folder. If you're seeing multiple .gguf files, ensure you're selecting the one with the correct quantization type (i2_s, tl1, etc.) that matches your setup. Also, make sure the model was downloaded with the --local-dir flag and not just as a repo clone. If in doubt, rerun the setup using setup_env.py, which automatically prepares the model in the expected structure for inference.

Replies: 1 comment

Comment options

You must be logged in to vote
0 replies
Answer selected by pawpatrolrockie
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
2 participants