-
Notifications
You must be signed in to change notification settings - Fork 5.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add evaluation data to the LLMs docs #4225
Comments
OpenHands started fixing the issue! You can monitor the progress here. |
OpenHands started fixing the issue! You can monitor the progress here. |
OpenHands started fixing the issue! You can monitor the progress here. |
OpenHands started fixing the issue! You can monitor the progress here. |
OpenHands started fixing the issue! You can monitor the progress here. |
The workflow to fix this issue encountered an error. Please check the workflow logs for more information. |
Co-authored-by: Graham Neubig <[email protected]>
What problem or use case are you trying to solve?
The blog article available at https://www.all-hands.dev/blog/evaluation-of-llms-as-coding-agents-on-swe-bench-at-30x-speed describes the quality and cost of various language models.
We should summarize this information, and add it to docs/modules/usage/llms/llms.md in order to make it easier for users to know which models they are recommended to use. We can also link back to the article for people who want to see more comprehensive information.
The text was updated successfully, but these errors were encountered: