Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Can we use the matrix factorization model locally by downloading it to our local? #39

Open
nileshm11 opened this issue Jul 31, 2024 · 2 comments

Comments

@nileshm11
Copy link

nileshm11 commented Jul 31, 2024

I am using a matrix factorization (mf) router in a RAG application and want to download the MF model to my local system. Is this possible? Could you also explain how it works internally? Additionally, can we change the embedding model used by the MF router?

@iojw
Copy link
Collaborator

iojw commented Aug 3, 2024

Yes, it is possible to download this from HuggingFace! https://huggingface.co/routellm/mf_gpt4_augmented. However, note that you will still have to generate embeddings for OpenAI as input for the time being - @thwu1 is working on adding support for other embedding models in #17!

Please refer to the paper for specific details for how it works: https://arxiv.org/abs/2406.18665

@nileshm11
Copy link
Author

Could you please provide more details or share the implementation code for integrating this into a local system? Specifically, I'm interested in how to set up and use the model, including any necessary configurations and dependencies.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants