Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Supervised fine tuning (SFT) of an LLM using HuggingFace, no availabel files for LoRA #457

Open
vietvo89 opened this issue Jul 30, 2024 · 0 comments

Comments

@vietvo89
Copy link

Hi @NielsRogge

I tried to follow your notebook. But after trainer.train() and trainer.save_state() steps, I can't find where the LoRA is saved. I expect it should be in 'data/zephyr-7b-sft-lora' but I just found these files and I think they are not LoRA.
Screenshot from 2024-07-30 16-25-38

At inference step, I tried to load the trained model with LoRA, I got an error:

OSError: data[/zephyr-7b-sft-lora](http://localhost:8888/zephyr-7b-sft-lora) does not appear to have a file named config.json. Checkout 'https://huggingface.co/data/zephyr-7b-sft-lora/tree/None' for available files.

I do not know what's wrong. Can you help to resolve this problem?

Thanks

@vietvo89 vietvo89 changed the title Supervised fine tuning (SFT) of an LLM using HuggingFace Supervised fine tuning (SFT) of an LLM using HuggingFace, no availabel files for LoRA Jul 30, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant