Replies: 1 comment
-
so... is it possible in such way serve model? |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I have followed the steps on Unsloth official notebook Alpaca + Llama-3 8b full example and finetuned a llama 3 8B model and I wanted to serve it using vllm? However it does not seem to work.
This is the command I used for serving the local model, with "/content/merged_llama3" being the directory that contains all model files:
which returns an error:
I dont think I ever need to provide quantization method as it should be written in the config file, it should be a mistake reading all those files. In addition, I did save the model and pushed it to the hub using the given codes in the Unsloth notebook?
my model files:
what went wrong?
Beta Was this translation helpful? Give feedback.
All reactions