Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

How to handle max token usage? #799

Open
Rodrigo-JM opened this issue Jan 28, 2025 · 1 comment
Open

How to handle max token usage? #799

Rodrigo-JM opened this issue Jan 28, 2025 · 1 comment
Assignees
Labels
more info More information required question Further information is requested

Comments

@Rodrigo-JM
Copy link

I am using a pretty standard rag approach but it quickly goes over the token limit for the o-mini model.

Need help since I didn't find any reference for this in the docs.

In other libraries such as LangGraph we have control over the messages in the course of a run.

Is there something similar here? Do I have to use graphs? How are you guys handling this?

@sydney-runkle sydney-runkle added the question Further information is requested label Jan 29, 2025
@sydney-runkle sydney-runkle self-assigned this Jan 29, 2025
@sydney-runkle
Copy link
Member

Hi @Rodrigo-JM, are you using UsageLimits?

@sydney-runkle sydney-runkle added the more info More information required label Jan 30, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
more info More information required question Further information is requested
Projects
None yet
Development

No branches or pull requests

2 participants