You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
After the jobs are confirmed to work on SILNLP, they are moved to SF/Serval. To ensure that the quality has not decreased in the transition (or with successive builds), there need to be some updates to Serval/Machine.py.
The overall proposal would be that when the final configuration is determined by EITL, SILNLP would perform a run and output in a Json file:
A set of verses for validation
A description of the training setup
The bleu score (or set of bleu scores per book/verse, etc., or CHRF++, etc.)
Then, this Json file would be uploaded to SF. SF would then take the validation split verses and create an engine in Serval to be used for performing this type of evaluation run. When a new build is made, SF would specify the validation verses explicitly (and early stopping or a number of training steps if desired). Serval would perform the build and return the Bleu CHRF++ scores per validation verses.
SF would take these scores and compare them against the known good set and send an email (or other type of alert) to the EITL team to address the issue.
The text was updated successfully, but these errors were encountered:
This idea is detailed here.
After the jobs are confirmed to work on SILNLP, they are moved to SF/Serval. To ensure that the quality has not decreased in the transition (or with successive builds), there need to be some updates to Serval/Machine.py.
The overall proposal would be that when the final configuration is determined by EITL, SILNLP would perform a run and output in a Json file:
Then, this Json file would be uploaded to SF. SF would then take the validation split verses and create an engine in Serval to be used for performing this type of evaluation run. When a new build is made, SF would specify the validation verses explicitly (and early stopping or a number of training steps if desired). Serval would perform the build and return the Bleu CHRF++ scores per validation verses.
SF would take these scores and compare them against the known good set and send an email (or other type of alert) to the EITL team to address the issue.
The text was updated successfully, but these errors were encountered: