You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
The issue you're encountering, where the service defaults to a model with a suffix like -qUQeFaHA instead of the specified glm4-9b, might be related to how the available models are detected and selected in the codebase.
Model Detection and Selection: The get_default_llm() function is responsible for selecting the default model. It checks if the DEFAULT_LLM_MODEL is in the list of available models. If not found, it logs a warning and defaults to the first available model [1].
Model Suffix: The suffix in model names like -qUQeFaHA might be generated or assigned during the model detection process, possibly to differentiate between similar models or versions. However, the exact mechanism for generating this suffix is not detailed in the available information [2].
Debugging Steps:
Ensure that the DEFAULT_LLM_MODEL is correctly specified in your configuration files and matches exactly with the model names listed in your available models.
Check the get_config_models function to see how models are being detected and if there are any discrepancies in the model names or configurations [3].
Verify if the model detection logic is correctly identifying the models from your configured platforms, such as Xinference.
Potential Solutions:
Double-check your configuration files to ensure there are no typos or mismatches in the model names.
If the suffix is being added dynamically, investigate the model detection logic to understand why this is happening and adjust the configuration or code as needed.
If these steps do not resolve the issue, further investigation into the model detection and configuration logic might be necessary to pinpoint the source of the problem.
问题描述 / Problem Description
在模型配置文件中已经设置了期望的glm4-9b作为LLM模型,但启动服务后一直报warning称找不到,且默认用了一个带后缀的9b模型,如图所示。
但是使用apipost是可以成功调到对应接口的。
debug时,在warning出现的对应脚本设置断点,但发现代码根本不会执行到对应脚本,也依然能弹出warning。
求助大家是否遇到了类似问题,如何解决?
复现问题的步骤 / Steps to Reproduce
依照源码部署md,start -a后出现相关问题
预期的结果 / Expected Result
没有warning,且可以在后台查到正确的模型调用记录
实际结果 / Actual Result
实际上期望的模型并没有被正确调用,且也不知道项目中自动找到的带后缀模型是从哪里调出来的
环境信息 / Environment Information
附加信息 / Additional Information
添加与问题相关的任何其他信息 / Add any other information related to the issue.
The text was updated successfully, but these errors were encountered: