You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
And what is the max value of max_model_len for DeepSeek-V2-Chat?
from transformers import AutoTokenizer
from vllm import LLM, SamplingParams
max_model_len, tp_size = 8192, 8
The text was updated successfully, but these errors were encountered:
ybdesire
changed the title
Why max_model_len only 8192 when inferencing with vLLM?
Why max_model_len only 8192 when inferencing with vLLM for DeepSeek-V2-Chat?
Jul 18, 2024
And what is the max value of
max_model_len
for DeepSeek-V2-Chat?The text was updated successfully, but these errors were encountered: