Skip to content

GroundZeros/Tricks-of-Semi-supervisedDeepLeanring-Pytorch

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

31 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Tricks of Semi-supervised Deep Leanring --Pytorch

The repository implements following semi-supervised deep learning methods:

  • PseudoLabel 2013: The Simple and Efficient Semi-Supervised Learning Method for Deep Neural Networks (ICMLW 2013)

  • PI&Tempens: Temporal Ensembling for Semi-Supervised Learning (ICLR 2017)

  • MeanTeacher: Mean Teachers are better Role Models (NIPS 2017)

  • VAT: Virtual Adversarial Training: A Regularization Method for Supervised and Semi-supervised Learning (TPAMI 2018)

  • ICT: Interpolation Consistency Training for Semi-supervised Learning (IJCAI 2019)

  • MixMatch: A Holistic Approach to Semi-Supervised Learning (NIPS 2019)

  • FixMatch: Simplifying Semi-Supervised Learning with Consistency and Confidence (2020)

This repository was created for my blog 半监督深度学习训练和实现小Tricks. Therefore the hyper-parameters are set for fair comparision, rather than performance.

The environment:

  • Ubuntu 16.04 + CUDA 9.0

  • Python 3.6.5:: Anaconda

  • PyTorch 0.4.1 and torchvision 0.2.1

To run the code:

The script run.sh includes some examples. You can try it as follow:

bash run.sh [gpu_id]

Some experimental results:

I haven't run all models in this repository. Some results of this repo. are shown in results directory. And the following results came from this repository and the old codes which this repo. built on.

The following table shows the error rates of the CIFAR10 experiment with 4000 labeled training samples. The parameter settings are the same with the examples in run.sh.

iPseudoLabel2013 ePseudoLabel2013 MeanTeacher MixMatch iFixMatch
orginal 12.31 6.24 4.26
v1 20.03 12.03 10.59 6.70 6.63
v2 15.82 10.82 9.46 6.89 6.44
iTempens eTempens PI ICT* VAT
orginal 12.16 13.20 7.29 11.36
v1 10.98 10.74 14.11 7.12 13.84
v2 13.53 10.24 12.89 6.74 12.67
eMixPseudoLabelv1 eMixPseudoLabelv2
soft 7.30 7.08
hard 7.33 7.20

Notes:

  • The MeanTeacher is the first model of the repository. So the hyper-parameters actually have been tuned for MeanTeacher.

  • My ICT is different from original one. The main difference is the unsupervised loss for unlabeled data.

  • eMixPseudoLabel is ePseudoLabel2013 with MixUp.

  • Instead of KL Divergence, my VAT uses MSE which bring more performance improvement.

About

PseudoLabel 2013, VAT, PI model, Tempens, MeanTeacher, ICT, MixMatch, FixMatch

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 94.1%
  • Shell 5.9%