diff --git a/optimum/habana/transformers/trainer.py b/optimum/habana/transformers/trainer.py index ee1427227d..4f22bdde37 100644 --- a/optimum/habana/transformers/trainer.py +++ b/optimum/habana/transformers/trainer.py @@ -1173,7 +1173,11 @@ def _maybe_log_save_evaluate(self, tr_loss, _grad_norm, model, trial, epoch, ign if is_accelerate_available() and self.accelerator.distributed_type == GaudiDistributedType.DEEPSPEED: grad_norm = model.get_global_grad_norm() else: - grad_norm = _grad_norm.item() if _grad_norm is not None else None + grad_norm = ( + _grad_norm.item() + if (_grad_norm is not None and self.accelerator.distributed_type != GaudiDistributedType.FSDP) + else None + ) if grad_norm is not None: logs["grad_norm"] = grad_norm