Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

How to use LLM deployed locally? #95

Open
deeper-coder opened this issue Oct 29, 2024 · 2 comments
Open

How to use LLM deployed locally? #95

deeper-coder opened this issue Oct 29, 2024 · 2 comments

Comments

@deeper-coder
Copy link

deeper-coder commented Oct 29, 2024

Your current code uses OpenAI’s API key to access the LLM service by default. I’d like to switch it to use a local LLM, which I have deployed through LLaMA-Factory and is accessible via a local API, for example, at http://localhost:7788/v1/. Could you guide me on how to make this adjustment? Thank you!

Upvote & Fund

  • We're using Polar.sh so you can upvote and help fund this issue.
  • We receive the funding once the issue is completed & confirmed by you.
  • Thank you in advance for helping prioritize & fund our backlog.
Fund with Polar
@kyegomez
Copy link
Owner

@deeper-coder we need function calling, if you can get a function calling model to work reliably then it will work. But you need a class with a run(task: str) or __call__(task: str) method to integrate into the ToTAgent class

@deeper-coder
Copy link
Author

I plan to use llama3 70B and I noticed that in the OpenAIFunctionCaller class, you’ve implemented the run method as shown in the image. So, can I achieve my desired functionality by passing base_url = "http://localhost:7788/v1/" in **kwargs?
image

@deeper-coder we need function calling, if you can get a function calling model to work reliably then it will work. But you need a class with a run(task: str) or __call__(task: str) method to integrate into the ToTAgent class

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants