You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
This issue has been automatically marked as inactive due to lack of recent activity. Should you believe it remains unresolved and warrants attention, kindly leave a comment on this thread.
Model Series
Qwen2.5
What are the models used?
Qwen2.5-7B-Instruct
What is the scenario where the problem happened?
deployment with vllm, tool calling with ollama
Is this a known issue?
Information about environment
OS: Linux
Python: 3.10.12
PyTorch: 2.4.0+cpu
vLLM: v0.6.1.post2
Ollama: 0.3.10
Log output
Description
Ollama
ollama request:
curl http://localhost:11434/v1/chat/completions -d '{
"model": "qwen2.5",
"temperature": 0.01,
"max_tokens": 2000,
"stream": false,
"messages": [
{
"role": "user",
"content": "系统状态"
}
],
"tools": [
{
"type": "function",
"function": {
"name": "cInBMXzQGq58",
"description": "系统状态方面",
"parameters": {
"type": "object",
"properties": {},
"required": []
}
}
}
],
"tool_choice": "auto"
}'
ollama response:
{
"id": "chatcmpl-724",
"object": "chat.completion",
"created": 1728697253,
"model": "qwen2.5",
"system_fingerprint": "fp_ollama",
"choices": [
{
"index": 0,
"message": {
"role": "assistant",
"content": "",
"tool_calls": [
{
"id": "call_l565rajv",
"type": "function",
"function": {
"name": "cInBMXzQz58",
"arguments": "{}"
}
}
]
},
"finish_reason": "tool_calls"
}
],
"usage": {
"prompt_tokens": 151,
"completion_tokens": 23,
"total_tokens": 174
}
}
vLLM
vLLM request:
vLLM response:
The text was updated successfully, but these errors were encountered: