-
Notifications
You must be signed in to change notification settings - Fork 5
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Bug in pseudo-labelling code? #3
Comments
I am also confused by the
This coefficient does not appear in the original paper. |
One more question: For self training, it seems both labeled and unlabeled data are used for the KL divergence between teacher and student? The original paper says only the unlabeled data is used to compute the KLD. |
Hello, for the first question, the code for pseudo-label is from this PyTorch repo which is a re-implementation from this official Tensorflow implementation from Google. From their comment: As for the As for the About self-training: Thanks for pointing this out. It is a typo in the paper, I did indeed use both labeled and unlabeled data for self-training. |
Thanks for the helpful response! Could you point me to any paper that uses this specific variant of this pseudo-labelling loss? |
I am confused by the implementation of pseudo-labelling in this library (lib/algs/pseudo_label.py). Especially, the forward() has:
I am confused why when computing
p_target
, thegt_mask
is multiplied by10
? What is meaning of10
here?Also, I believe the
lt_mask
means the examples with max probability smaller than threshold and thus should be ignored when computing the loss. However, thep_target
has the+ lt_mask[:,None] * y_probs
.This seems to be different from what is described in the paper. If you are implementing a variant of pseudo-labelling loss function, could you point me to that paper?
The text was updated successfully, but these errors were encountered: