Skip to content

Commit

Permalink
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
revert decorator in UT
Browse files Browse the repository at this point in the history
tocean committed Feb 1, 2024
1 parent 67c0a3f commit 2150f64
Showing 1 changed file with 2 additions and 2 deletions.
4 changes: 2 additions & 2 deletions tests/fsdp/test_fsdp_distributed.py
Original file line number Diff line number Diff line change
@@ -9,7 +9,7 @@
import torch.distributed as dist
from torch.testing._internal.common_distributed import MultiProcessTestCase, skip_if_lt_x_gpu, requires_nccl

#from tests.helper import decorator
from tests.helper import decorator
from msamp.fsdp import FsdpReplacer, FP8FullyShardedDataParallel


@@ -37,7 +37,7 @@ def world_size(self):

@requires_nccl()
@skip_if_lt_x_gpu(2)
#@decorator.cuda_test
@decorator.cuda_test
def test_fp8_fsdp(self):
"""Test forward and backward functionality in FP8 FSDP."""
rank = self.rank

0 comments on commit 2150f64

Please sign in to comment.