-
Notifications
You must be signed in to change notification settings - Fork 35
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
How to Perform Inference with Fine-Tuned VideoLLaMA3? #32
Comments
Maybe step 3 can help to finetune your own data in VideoLLaMA 3: "For finetuneing, --model_path is the path to the converted checkpoint as described in step 2." I also fallow VideoLLaMA series work. We have a wechat group to help each other sloving issues. You can add my wechat number(19357260600) or E-mail([email protected]) to talk with us. |
Yes, I followed the fine-tuning instructions to convert the model checkpoint and fine-tuned the model using this checkpoint. Do I need to copy same configuration and process files from origin videollama3-2B folder? |
I'm not sure yet, I'm deploying videollama3-7B, ready to train and inference my vertical video data. |
Hi @JasonYeh0111 , thanks for your interests! For trying your local model, you can follow videollama3/infer.py and you just need to replace the |
Thanks for your great work! I have a question: After fine-tuning VideoLLaMA3-2B, how can I use the model for inference? The fine-tuned model has different weights compared to the original VideoLLaMA3-2B, as it includes a vision encoder. Should I use the same configuration file as the original VideoLLaMA3-2B?
The text was updated successfully, but these errors were encountered: