From 395b52d65419a9a71054a2e2ee207599ada45f71 Mon Sep 17 00:00:00 2001 From: Cheng Penghui Date: Mon, 5 Aug 2024 07:13:46 +0000 Subject: [PATCH] Remove unnecessary cases from skip list Signed-off-by: Cheng Penghui --- test/xpu/run_test_with_skip.py | 22 ---------------------- test/xpu/test_modules_xpu.py | 2 +- 2 files changed, 1 insertion(+), 23 deletions(-) diff --git a/test/xpu/run_test_with_skip.py b/test/xpu/run_test_with_skip.py index e7c34283e..35d675650 100644 --- a/test/xpu/run_test_with_skip.py +++ b/test/xpu/run_test_with_skip.py @@ -1134,37 +1134,18 @@ def launch_test(test_case, skip_list=None, exe_list=None): "test_save_load_nn_TransformerEncoder_eval_mode_xpu_float64", "test_save_load_nn_TransformerEncoder_train_mode_xpu_float64", "test_save_load_nn_Transformer_xpu_float64", - # AssertionError: Tensor-likes are not close! - "test_cpu_gpu_parity_nn_ConvTranspose3d_xpu_complex32", # Unexpected success: "test_cpu_gpu_parity_nn_ConvTranspose2d_xpu_complex32", "test_cpu_gpu_parity_nn_ConvTranspose1d_xpu_complex32", "test_memory_format_nn_AvgPool2d_xpu_float32", "test_memory_format_nn_AvgPool2d_xpu_float64", - # AssertionError: False is not true - "test_memory_format_nn_Conv2d_xpu_float64", - "test_memory_format_nn_ConvTranspose2d_xpu_float64", - "test_memory_format_nn_LazyConv2d_xpu_float64", - "test_memory_format_nn_LazyConvTranspose2d_xpu_float64", # CPU fallback could not cover these # CUDA xfails # Failed: Unexpected success "test_memory_format_nn_AdaptiveAvgPool2d_xpu_float32", "test_memory_format_nn_AdaptiveAvgPool2d_xpu_float64", # CPU fallback fails - # AssertionError: False is not true - "test_memory_format_nn_ReflectionPad3d_xpu_float32", - "test_memory_format_nn_ReflectionPad3d_xpu_float64", - "test_memory_format_nn_ReplicationPad2d_xpu_float32", - "test_memory_format_nn_ReplicationPad2d_xpu_float64", - "test_memory_format_nn_ReplicationPad3d_xpu_float32", - "test_memory_format_nn_ReplicationPad3d_xpu_float64", - # CPU fallback fails # RuntimeError: view size is not compatible with input tensor's size and stride (at least one dimension spans across two contiguous subspaces). Use .reshape(...) instead. - "test_memory_format_nn_GroupNorm_xpu_bfloat16", - "test_memory_format_nn_GroupNorm_xpu_float16", - "test_memory_format_nn_GroupNorm_xpu_float32", - "test_memory_format_nn_GroupNorm_xpu_float64", "test_save_load_nn_GRU_eval_mode_xpu_float32", "test_save_load_nn_GRUCell_xpu_float32", "test_save_load_nn_GRU_train_mode_xpu_float32", @@ -1189,9 +1170,6 @@ def launch_test(test_case, skip_list=None, exe_list=None): "test_non_contiguous_tensors_nn_GRUCell_xpu_float32", "test_non_contiguous_tensors_nn_GRU_eval_mode_xpu_float32", "test_non_contiguous_tensors_nn_GRU_train_mode_xpu_float32", - # Reflection_pad2d doesn't support channel last, CUDA skipped too. - "test_memory_format_nn_ReflectionPad2d_xpu_float32", - "test_memory_format_nn_ReflectionPad2d_xpu_float64", # AssertionError: False is not true "test_to_nn_BatchNorm1d_eval_mode_swap_True_set_grad_True_xpu_float32", "test_to_nn_BatchNorm1d_train_mode_swap_True_set_grad_True_xpu_float32", diff --git a/test/xpu/test_modules_xpu.py b/test/xpu/test_modules_xpu.py index 49cac1f6e..06bf8d114 100644 --- a/test/xpu/test_modules_xpu.py +++ b/test/xpu/test_modules_xpu.py @@ -93,7 +93,7 @@ def _test_to(self, device, dtype, module_info, training, swap, set_grad): devices = ['xpu'] dtypes = module_info.dtypes module_inputs = module_info.module_inputs_func(module_info, device=device, dtype=dtype, - requires_grad=False, training=training) + requires_grad=False, training=training) torch.__future__.set_swap_module_params_on_conversion(swap) for module_input in module_inputs: