Skip to content

Deep Learning framework that trains a single model using multiple, heterogeneous datasets leveraging parallel transfer, strictly enforcing feature generalization and even preventing overfitting

License

Notifications You must be signed in to change notification settings

eXascaleInfolab/typhon

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

2 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Typhon

Typhon is a new Deep Learning framework that trains a single model using multiple, heterogeneous datasets leveraging parallel transfer. This aims to improve the performance of Deep Learning methods in critical applications afflicted by data scarcity, such as computer-aided diagnosis for cancer detection, where large datasets are rare or unfeasible but many smaller datasets may be available. The key idea is to assemble sufficient data for training deep models by selecting a set of multiple, potentially smaller and heterogeneous datasets, as long as they all exhibit similar visual features, such as common with medical imaging applications. The Typhon model architecture is composed of a single Feature Extractor and multiple Decision Makers, in sequence but explicitly separated. The Feature Extractor is trained using all datasets with a focus on producing generic features which are useful across all datasets. The Decision Makers are each paired with a different dataset, and specialized to take decisions based on the output of the Feature Extractor. Our training method is based on the concept of parallel transfer: on each epoch, we train on just one batch from each dataset in turn. This is done by pairing the correct Decision Maker on top of the shared Feature Extractor, then training the resulting model end-to-end on the data batch using classical methods. The actual design is inherently more complex, as we had to overcome a set of major challenges such as dataset imbalance, moving target, catastrophic forgetting and issues with initialization viability. Once made viable however, this methods excels at strictly enforcing feature generalization and even preventing overfitting. Our code for the experiments presented in our publication are available in this repository.

About

Deep Learning framework that trains a single model using multiple, heterogeneous datasets leveraging parallel transfer, strictly enforcing feature generalization and even preventing overfitting

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages