LoRA support #65
11415142513152119
started this conversation in
Ideas
Replies: 2 comments 4 replies
-
ExLlamaV2 doesn't currently support LoRAs. Support is coming soon, though. After that there may need to be some minor changes to TGW to support whatever the API ends up looking like. |
Beta Was this translation helpful? Give feedback.
3 replies
-
@turboderp qq: I see that exllamav2 supports Multi-Lora, how is that different from https://github.com/punica-ai/punica or https://github.com/S-LoRA/S-LoRA? Would I be able to use exllama to hot-swap LoRA's at request time? |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Would LoRA support need to be implemented on this end or on the backend end?
Right now trying to load them on Ooba's results in either "AttributeError: 'Exllamav2Model' object has no attribute 'dtype'" on ExLlamav2 or "ValueError: Target modules ['k_proj', 'q_proj', 'up_proj', 'o_proj', 'gate_proj', 'down_proj', 'v_proj'] not found in the base model. Please check the target modules and try again." on ExLlamav2_HF.
Beta Was this translation helpful? Give feedback.
All reactions