-
Notifications
You must be signed in to change notification settings - Fork 451
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Feature]: support LlavaForConditionalGeneration with turbomind inference #2710
Conversation
Please resolve the linting error by : pip install pre-commit
cd lmdeploy # the root directory of lmdeploy repo
pre-commit install
pre-commit run --all-files |
bf61682
to
20c7476
Compare
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Please resolve the conflicts
- fix tune attribute error - add chinese llava doc
3383c6e
to
48d1a5c
Compare
从config.json 里面读参数可能不是特别好,尤其是对于这种融合了多个模型的结构,很多参数都省略了,强行加默认值会增加出错的风险,之后可以另外提交PR采用更好的方式。 |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Tested OK with llava-interleave-qwen-7b-hf
|
…ence (InternLM#2710) * feat: support llava_qwen2 for fp16 and awq * update generate gemm config script for VLM * lint: fix lint warning * doc: presenting the usage in the user guide * resolve conflict issue and refactor for better design * fix and doc: - fix tune attribute error - add chinese llava doc * keep LlavaLlamaForCausalLM/LlavaMistralForCausalLM to llama * fix attn_bias default value
Motivation
Modification
Use cases (Optional)
lmdeploy serve api_server llava_hf/llava_interleave_qwen2_7b_hf