diff --git a/axonn/tests/test_intra_layer_conv.py b/axonn/tests/test_intra_layer_conv.py index f6779b5..a304a6f 100644 --- a/axonn/tests/test_intra_layer_conv.py +++ b/axonn/tests/test_intra_layer_conv.py @@ -42,7 +42,6 @@ def norm_allclose(X, Y): @pytest.mark.parametrize("easy_tp", [True, False]) @pytest.mark.parametrize("bias", [True, False]) @pytest.mark.parametrize("device", ["cuda", "cpu"]) -@pytest.mark.xfail(reason="Allow to fail as convolutions are not deterministic") def test_fw_pass(G_intra_r, G_intra_c, G_intra_d, B, H, W, C, easy_tp, bias, device): # These tests are in fp-32 torch.manual_seed(42) @@ -136,7 +135,6 @@ def test_fw_pass(G_intra_r, G_intra_c, G_intra_d, B, H, W, C, easy_tp, bias, dev @pytest.mark.parametrize("bias", [True, False]) @pytest.mark.parametrize("device", ["cuda", "cpu"]) @pytest.mark.parametrize("comm_opt_level", [0, 3]) -@pytest.mark.xfail(reason="Allow to fail as convolutions are not deterministic") def test_bw_pass( G_intra_r, G_intra_c,