You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Add support for users to configure a secondary, lightweight LLM alongside their main LLM. This auxiliary LLM would handle quick, repetitive tasks where using their primary (potentially expensive/slower) LLM would be overkill.
If no auxiliary LLM is configured, we'll fall back to some hardcoded solutions or main LLM, as we already have.
Motivation
Users configure their own LLMs in our project, which we recommend to be some powerful models. Using these for simple tasks like generating branch names or checking for agent loops is unnecessarily expensive and slow. We could:
give users the option to use a smaller, faster LLM for simple tasks (not agent tasks)
save cost / reduce latency/rate limits on the main LLM
Use Cases
The resolver has currently a couple of cases which are an example of things we may use such LLM for (guess success, review)
In addition:
Workspace Summaries
generate quick summaries for workspace.zip downloads
example: "workspace_add-user-auth.zip" instead of just "workspace.zip"
fallback: Use workspace.zip, or branch name if a repo exists, or timestamp
I think this may be somewhat related to the not diamond issue #4184, but maybe slight different, bc you were thinking about using it for simple agent tasks.
I think this could be achievable by maintaining an LLM instance inside agent_controller? But allowing user to manually set TWO LLM config could be too tricky - I guess we can just use 4o-mini for openai/, haiku for anthropic/ LLM key, and fall back to existing implementation for other providers?
This is needed even in the era of 200K Context, since terminal output + complex code calling logic, are also things that work against context length. Also as a practical concern, token = cost. Anything that compress context is good.
Summary
Add support for users to configure a secondary, lightweight LLM alongside their main LLM. This auxiliary LLM would handle quick, repetitive tasks where using their primary (potentially expensive/slower) LLM would be overkill.
If no auxiliary LLM is configured, we'll fall back to some hardcoded solutions or main LLM, as we already have.
Motivation
Users configure their own LLMs in our project, which we recommend to be some powerful models. Using these for simple tasks like generating branch names or checking for agent loops is unnecessarily expensive and slow. We could:
Use Cases
The resolver has currently a couple of cases which are an example of things we may use such LLM for (guess success, review)
In addition:
Workspace Summaries
Stuck-in-a-loop
Branch Name Creation
Alternatives to Consider
The text was updated successfully, but these errors were encountered: