Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

cannot run hg38_hyena_seqlen_warmup_reload #45

Open
sustcsonglin opened this issue Jan 14, 2024 · 2 comments
Open

cannot run hg38_hyena_seqlen_warmup_reload #45

sustcsonglin opened this issue Jan 14, 2024 · 2 comments

Comments

@sustcsonglin
Copy link

sustcsonglin commented Jan 14, 2024

Hi,

Thanks for your great work! I am trying to do hg38/hg38_hyena_seqlen_warmup_reload.yaml experiment. Got the following error msg:

image

I had some initial search on this issue and found this. I set monitor: test/loss and it still doesn't work. But i have no problem running 'g38/hg38_hyena.yaml'.

Do you have any insights on this issue? Is this related to the sentence length warmup callback? because I can run 'g38/hg38_hyena.yaml' without this callback. i am using pytorch_lightning v1.8.6

@sustcsonglin
Copy link
Author

After some inspections, I found the reason! link1 disables loading the test dataloader during training, but the monitor metric was not changed. Hope this configuration will be fixed

@sustcsonglin
Copy link
Author

Another question: does this configuration file consider the trainer.devices for setting accumulate_grad_batches? I didn't find sth like dividing by trainer.devices here

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant