Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

How to Perform Inference with Fine-Tuned VideoLLaMA3? #32

Open
JasonYeh0111 opened this issue Feb 20, 2025 · 4 comments
Open

How to Perform Inference with Fine-Tuned VideoLLaMA3? #32

JasonYeh0111 opened this issue Feb 20, 2025 · 4 comments

Comments

@JasonYeh0111
Copy link

Thanks for your great work! I have a question: After fine-tuning VideoLLaMA3-2B, how can I use the model for inference? The fine-tuned model has different weights compared to the original VideoLLaMA3-2B, as it includes a vision encoder. Should I use the same configuration file as the original VideoLLaMA3-2B?

@LiangMeng89
Copy link

Maybe step 3 can help to finetune your own data in VideoLLaMA 3: "For finetuneing, --model_path is the path to the converted checkpoint as described in step 2."

I also fallow VideoLLaMA series work. We have a wechat group to help each other sloving issues. You can add my wechat number(19357260600) or E-mail([email protected]) to talk with us.

@JasonYeh0111
Copy link
Author

Yes, I followed the fine-tuning instructions to convert the model checkpoint and fine-tuned the model using this checkpoint. Do I need to copy same configuration and process files from origin videollama3-2B folder?

@LiangMeng89
Copy link

Yes, I followed the fine-tuning instructions to convert the model checkpoint and fine-tuned the model using this checkpoint. Do I need to copy same configuration and process files from origin videollama3-2B folder?

I'm not sure yet, I'm deploying videollama3-7B, ready to train and inference my vertical video data.

@lkhl
Copy link
Collaborator

lkhl commented Feb 24, 2025

Hi @JasonYeh0111 , thanks for your interests!

For trying your local model, you can follow videollama3/infer.py and you just need to replace the model_path to the path of your checkpoint.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants