You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Deep learning is very resource intensive. A common trick to reduce compute needs is to lower the precision of your data (e.g., from float32 to float16). Alternatively, there are options to use automatic mixed precision via torch.amp. We should provide some guidelines on the impact of lowering precision (e.g., during training vs. during inference). Provide references/evidence to support these guidelines.
The text was updated successfully, but these errors were encountered:
qualiaMachine
changed the title
Guidance on using lower precision for deep learning
Guidance on using lower precision ("quantization") for deep learning
Aug 21, 2024
Deep learning is very resource intensive. A common trick to reduce compute needs is to lower the precision of your data (e.g., from float32 to float16). Alternatively, there are options to use automatic mixed precision via torch.amp. We should provide some guidelines on the impact of lowering precision (e.g., during training vs. during inference). Provide references/evidence to support these guidelines.
The text was updated successfully, but these errors were encountered: