Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

What did OLMo 1B converge to? #642

Open
sidereior opened this issue Jul 4, 2024 · 1 comment
Open

What did OLMo 1B converge to? #642

sidereior opened this issue Jul 4, 2024 · 1 comment
Labels
type/question An issue that's a question

Comments

@sidereior
Copy link

❓ The question

Hi all,
Thanks so much for this amazing repo. I'm training a 1B model from scratch and am just wondering what it the final loss converged to and what the final perplexity is. Thanks!

@sidereior sidereior added the type/question An issue that's a question label Jul 4, 2024
@AkshitaB
Copy link
Contributor

Hi, you can take a look at the training metrics here: https://wandb.ai/ai2-llm/OLMo-1B/reports/OLMo-1B--Vmlldzo2NzY1Njk1

The final perplexity was 10.815, and the final loss was 2.376

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
type/question An issue that's a question
Projects
None yet
Development

No branches or pull requests

2 participants