Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

System role problem running Gemma 2 on vLLM #1386

Closed
juanjuanignacio opened this issue Aug 5, 2024 · 5 comments
Closed

System role problem running Gemma 2 on vLLM #1386

juanjuanignacio opened this issue Aug 5, 2024 · 5 comments
Labels
support A request for help setting things up

Comments

@juanjuanignacio
Copy link

juanjuanignacio commented Aug 5, 2024

Hello,

In running chat ui and trying some models, with phi3 and llama i had no problem but when I run gemma2 in vllm Im not able to make any good api request,
in env.local:
{
"name": "google/gemma-2-2b-it",
"id": "google/gemma-2-2b-it",
"chatPromptTemplate": "{{#each messages}}{{#ifUser}}<start_of_turn>user\n{{#if @FIRST}}{{#if @root.preprompt}}{{@root.preprompt}}\n{{/if}}{{/if}}{{content}}<end_of_turn>\n<start_of_turn>model\n{{/ifUser}}{{#ifAssistant}}{{content}}<end_of_turn>\n{{/ifAssistant}}{{/each}}",
"parameters": {
"temperature": 0.1,
"top_p": 0.95,
"repetition_penalty": 1.2,
"top_k": 50,
"truncate": 1000,
"max_new_tokens": 2048,
"stop": ["<end_of_turn>"]
},
"endpoints": [
{
"type": "openai",
"baseURL": "http://127.0.0.1:8000/v1",

  }
]

}

and I always have the same response in vllm server:

ERROR 08-05 12:39:06 serving_chat.py:118] Error in applying chat template from request: System role not supported
INFO: 127.0.0.1:42142 - "POST /v1/chat/completions HTTP/1.1" 400 Bad Request

do someone know if I have to change and how do change the chat template or deactivate system role ? is it a vllm problem or a chat ui problem?

Thank U!

@nsarrazin nsarrazin added the support A request for help setting things up label Aug 8, 2024
@hieutrungdao
Copy link

Same issue :'(

@markVaykhansky
Copy link

+1

@llllvvuu
Copy link
Contributor

Kinda hacky but you can change {{ raise_exception('System role not supported') }} in the Gemma2 chat template to {%- set messages = messages[1:] %}. It works for me with that change.

But I think it would be nice to be able to omit the system message from chat-ui side. Looks like the relevant code is here:

let messages: Message[] = [
{
id: v4(),
from: "system",
content: values.preprompt ?? "",
createdAt: new Date(),
updatedAt: new Date(),
children: [],
ancestors: [],
},
];

@nsarrazin
Copy link
Collaborator

Opened an issue for a potential solution, feel free to tackle it if you want! 😄 #1432

@nsarrazin
Copy link
Collaborator

Issue should be solved, try adding "systemRoleSupported": false to your model config!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
support A request for help setting things up
Projects
None yet
Development

No branches or pull requests

5 participants