feat: support togetherAI via /completions
#2045
Open
+193
−4
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Adds tested TogetherAI support
There are two ways to use TogetherAI:
The obvious way is to override
OPENAI_API_KEY
andOPENAI_BASE_URL
-> this will set up TogetherAI as an OpenAI proxy server, similar to the OpenRouter documentation here: https://docs.letta.com/models/openai_proxyHowever, upon testing it seems like the TogetherAI function calling support is pretty limited / has poor performance, so we probably want to use TogetherAI via
/completions
routes (similar to how we connect to vLLM) instead. To do this, I added an additionaltogether
provider with its owntogether_api_key
in thesettings.py
. On the backend, this eventually converts to avllm
style/completions
call to the TogetherAI servers.Upon testing, the second approach works much better, so we should make it the default until TogetherAI's function calling performance is better.
To test
export TOGETHER_API_KEY=...