You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
There will only be one saved_tensors_hooks active, specifically for the most recent gradient_checkpoint_scope. So any of the earlier pack hooks will not be used, when there are multiple simultaneous gradient_checkpoint_scopes.
A solution is that we keep a global weak tensor key dictionary for all registered tensors of any gradient_checkpoint_scope, and in the pack hook, check that instead of the local.
It's currently maybe not so important, as this is a case we likely do not run into (yet; I guess).
The text was updated successfully, but these errors were encountered:
There will only be one
saved_tensors_hooks
active, specifically for the most recentgradient_checkpoint_scope
. So any of the earlier pack hooks will not be used, when there are multiple simultaneousgradient_checkpoint_scope
s.Example code:
A solution is that we keep a global weak tensor key dictionary for all registered tensors of any
gradient_checkpoint_scope
, and in the pack hook, check that instead of the local.It's currently maybe not so important, as this is a case we likely do not run into (yet; I guess).
The text was updated successfully, but these errors were encountered: