Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[WIP] Fetch new validation data for each calc_loss #279

Draft
wants to merge 1 commit into
base: master
Choose a base branch
from

Conversation

Sopel97
Copy link
Collaborator

@Sopel97 Sopel97 commented Dec 22, 2020

This PR makes calc_loss use a new validation set each time it's called (one can now limit the amount of validation data used per calc_loss call, so this makes sense). Since we're using relatively small validation sets for each calc_loss this should, in principle at least, make validation loss more informative when observed over multiple validation steps.

I'd like to get some feedback from @noobpwnftw on this change before it's merged.

I'd also like to get some feedback on how to handle the case when we couldn't fetch the requested amount of the data for validation. This might happen in extreme cases. Currently I made it continue with a warning as the training can proceed even without validation data, but there might be a better resolution to this problem.

@Sopel97 Sopel97 changed the title Fetch new validation data for each calc_loss [WIP] Fetch new validation data for each calc_loss Dec 24, 2020
@NightlyKing
Copy link
Collaborator

As I already said in discord: the loss now really jumps around like crazy. My idea was to only slowly introduce new positions to the validation set to get a more stable loss output which will help with automatic LR drops.

@Sopel97 Sopel97 added the question Further information is requested label Apr 3, 2021
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
question Further information is requested
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants