[Feature Request] Support Stochastic Weight Averaging (SWA) for improved stability #321
Open
1 task done
Labels
enhancement
New feature or request
🚀 Feature
Stochastic Weight Averaging (SWA) is a recently proposed technique can potentially help improve training stability in DRL. There is now a new implementation in
torchcontrib
. Quoting/paraphrasing from their page:See the PyTorch SWA page for more.
Motivation
SWA might help improve training stability as well as final reward in some DRL scenarios. It may also alleviate sensitivity to random-seed initialization.
Pitch
See above :)
Alternatives
No response
Additional context
See the PyTorch SWA page for more.
Checklist
The text was updated successfully, but these errors were encountered: