-
Notifications
You must be signed in to change notification settings - Fork 7
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Track model metadata #481
Comments
If we know the token limit for a specific model, we can
|
Some thoughts on this can be found here cc @snopoke @stephherbers @bderenzi |
Langfuse has a nice way of doing it. Under the Tracing tab -> Models. This might serve as inspiration |
fyi, this is currently changing in Langfuse as more detailed model price details are necessary (cached/video/audio tokens). will probably be released next week |
It can really help users out if OCS knows a few things about selected models so that we can build guardrails that will ultimately help lower frustration and make the platform for robust. See this thread as an example where it would have been useful if OCS had known what the model's token limit is.
Model metadata to track
Currently we store model token limits on the Experiment or Pipeline Node but the values are specific to the LLM model so should be stored along with the model name.
We must also allow users to create new 'models' via the UI e.g. fine tuned models.
The token limits should then be accessed through the LlmService objects.
The text was updated successfully, but these errors were encountered: