Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

How to Load OpenELM Pre-training Checkpoints using Hugging Face AutoModelForCausalLM ? #43

Open
jasonkrone opened this issue Aug 8, 2024 · 3 comments

Comments

@jasonkrone
Copy link

Hi there,

First, really admire the work on OpenELM! Thank you for making your models and code available.

Question regarding the pre-training checkpoints linked here: how can we convert these checkpoints into the format expected by AutoModelForCausalLM.from_pretrained?

I presume there's a script that was used for conversion of the final model weights into HF format, but I couldn't find it in the repo.

Would very much appreciate any help on this!

Best,
Jason

@jasonkrone jasonkrone changed the title How to Load Pre-training Checkpoints using Hugging Face AutoModelForCausalLM ? How to Load OpenELM Pre-training Checkpoints using Hugging Face AutoModelForCausalLM ? Aug 8, 2024
@a154377713
Copy link

I have also encountered the same problem. Do you have a solution?

@jasonkrone
Copy link
Author

I didn't wind up solving this but here's a reference that might be helpful https://github.com/foundation-model-stack/foundation-model-stack/blob/4349dacef63e86b6c1acdccb69b48fe562365bb2/fms/models/llama.py#L592

@jasonkrone jasonkrone reopened this Sep 3, 2024
@athrvkk
Copy link

athrvkk commented Oct 9, 2024

As a follow question - Are there any plans to push the model checkpoints to Huggingface hub for ease of access (like the Pythia suite of models)?
It would really help the NLP community!
Thanks in Advance!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants