Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

add_thinking #5220

Closed
wants to merge 3 commits into from
Closed

add_thinking #5220

wants to merge 3 commits into from

Conversation

Tomlili43
Copy link

End-user friendly description of the problem this fixes or functionality that this introduces
Agent thinking feature by thinking

  • Include this change in the Release Notes. If checked, you must provide an end-user friendly description for your change below

Give a summary of what the PR does, explaining any non-trivial design decisions
add thinking prompt and thinking front end UI.

image
image


Link of any specific issues this addresses

Copy link
Contributor

@neubig neubig left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Hi @Tomlili43 , thanks for the contribution!

In general, all prompts should be implemented in the Python backend, not the frontend. And if we add a new prompt we'll need to validate its effect on agent accuracy and cost.

If you're interested in adding this prompt, could you try adding it in the agenthub/codeact_agent directory? (you can look at the directory structure there to understand what needs to be done)

We would then need to validate the effect of this change on performance and cost. I'm a bit worried that adding this big prompt in the input and encouraging longer outputs would lead to significant increases in time/token cost, but we could try a few examples to see if this makes a big difference in accuracy!

@Tomlili43
Copy link
Author

@neubig yeap,

  1. added it in the agenthub/codeact_agent directory.
  2. for validation, it will cost so many tokens. To be honest , i couldn't afford it. Can you help with it ?

@neubig neubig added the run-eval-s Runs evaluation with 5 instances label Nov 23, 2024
@neubig
Copy link
Contributor

neubig commented Nov 23, 2024

Hey @Tomlili43 , we can try to do this, but it'd be good if you could try running a few examples first to see if it's working!

@Tomlili43
Copy link
Author

@neubig hi, validation == run cmd: ./evaluation/agent_bench/scripts/run_infer.sh eval_thinking_prompt_llm HEAD CodeActAgent 1
right ?

@neubig
Copy link
Contributor

neubig commented Nov 27, 2024

Yep, that looks right.

@mamoodi
Copy link
Collaborator

mamoodi commented Dec 23, 2024

@Tomlili43 is this something you still wish to pursue?

@mamoodi
Copy link
Collaborator

mamoodi commented Dec 30, 2024

Going to close this. Please let me know if this was still in progress and we can reopen.

@mamoodi mamoodi closed this Dec 30, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
run-eval-s Runs evaluation with 5 instances
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants