pre-training masked language modeling accuracy #182
-
Hi all, Thank you for the amazing work. In your paper, the performance on various downstream tasks is noted. I am curious to find out about the pre-training masked language modeling accuracy. I am asking because I'm looking to learn more about peptide sequences using a much smaller model, and am curious to know what the pre-training accuracy benchmarks are. Thank you for your time and consideration. Regards, |
Beta Was this translation helpful? Give feedback.
Replies: 2 comments
-
Hi Vignesh, thanks for your interest! |
Beta Was this translation helpful? Give feedback.
-
Thank you for very much for the information. Regards, |
Beta Was this translation helpful? Give feedback.
Hi Vignesh, thanks for your interest!
We haven't really monitored accuracy during training. The pnas paper you link discusses our main metric ECE at length, which is directly related to the objective being minimized in MLM training.
Given the pre-trained models, it would be easy to compute accuracy on your dataset eg by masking out single positions one by one, or repeated random masking 15% in the way the model was trained.