-
Notifications
You must be signed in to change notification settings - Fork 47
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Finetuning or training ? #14
Comments
Hi @trmoro ! I contacted @bofenghuang to see if he could show me how to use finetune.py by giving me a demo with the instruction: "Donne moi le top 10 du ballon d'or 2016" and outputting the top 10. And all this while showing me the changes in the .json file. In short, a mini tuto of how to do it (in video). I'm waiting for his answer, but if you succeeded, could you make a screen recording and send it to me? I would like to have the steps of the fine-tuning that would help me a lot! |
I say this because I have a Macbook M1 and I don't have CUDA with NVIDIA. If you can also explain me the steps to do and the prerequisite I would be extremely grateful! |
Hi @trmoro! Have you checked the loss curve during fine-tuning? This can provide insight into whether the model is learning and improving. Also, how much training data did you use? Regarding hyperparameters, you might want to try using more aggressive values for parameters related to LoRA. This could potentially improve the model's performance. |
Hi @bofenghuang For the training, I did 150 epochs. The loss curve is decreasing fastly after epoch 25 (loss = 0.4 at epoch 25, loss =0.09 at epoch 50) I try with modified LoRA parameters. Here is the parameters I used before : python3 finetune.py |
My issue may be basic, but I've try finetuning vigogne with custom french instructions, but when i'm using this model, I never get the responses that I wrote.
So, I wondering if I'm using finetuning right or if I have to use a different method to train the model.
The text was updated successfully, but these errors were encountered: