LLaVA with GGUF files (and --mmproj argument)? #8341
Unanswered
benmayersohn
asked this question in
Q&A
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hi everyone! I'm coming from
llama.cpp
which currently doesn't support serving multi-modal models (it was removed and hopefully it will return soon).I'm used to the GGUF format, which is only recently supported by vLLM. In
llama.cpp
you would do something like this:where
${MMPROJ}
is a multi-modal projector that aligns the vision and text data (per this overview)${LLAVA_MODEL}
is the Llava model itselfand both are two distinct GGUF files. Does anyone know what the equivalent command/arguments be when serving via
python -m vllm.entrypoints.openai.api_server
?Thanks!
Beta Was this translation helpful? Give feedback.
All reactions