From 0d2a73ea83068ddeaa7bba983a89376661c36f83 Mon Sep 17 00:00:00 2001 From: Shucai Xiao Date: Thu, 25 Jan 2024 14:29:17 -0800 Subject: [PATCH] refine chain-dot.mlir --- test/TritonGPU/chain-dot.mlir | 9 ++++++--- 1 file changed, 6 insertions(+), 3 deletions(-) diff --git a/test/TritonGPU/chain-dot.mlir b/test/TritonGPU/chain-dot.mlir index f6b54aacbe36..e93e80ea9151 100644 --- a/test/TritonGPU/chain-dot.mlir +++ b/test/TritonGPU/chain-dot.mlir @@ -1,6 +1,7 @@ // RUN: triton-opt %s -split-input-file --tritonamdgpu-accelerate-matmul=arch-generation-name=gfx90a | FileCheck --check-prefixes=CHECK %s -// CHECK: #mfma = #triton_gpu.mfma<{version = 2.0, warpsPerCTA = [4, 1], instrShape = [32, 32], isTransposed = true}> +// CHECK: #mfma +// CHECK-SAME: isTransposed = true #blocked = #triton_gpu.blocked<{sizePerThread = [1, 4], threadsPerWarp = [8, 8], warpsPerCTA = [4, 1], order = [1, 0], CTAsPerCGA = [1, 1], CTASplitNum = [1, 1], CTAOrder = [0, 1]}> #blocked1 = #triton_gpu.blocked<{sizePerThread = [2, 2], threadsPerWarp = [4, 16], warpsPerCTA = [4, 1], order = [1, 0], CTAsPerCGA = [1, 1], CTASplitNum = [1, 1], CTAOrder = [1, 0]}> @@ -64,7 +65,8 @@ module attributes {"triton_gpu.compute-capability" = 0 : i32, "triton_gpu.num-ct } -// CHECK: #mfma1 = #triton_gpu.mfma<{version = 2.0, warpsPerCTA = [4, 1], instrShape = [32, 32], isTransposed = false}> +// CHECK: #mfma1 +// CHECK-SAME: isTransposed = false #blocked2 = #triton_gpu.blocked<{sizePerThread = [1, 4], threadsPerWarp = [8, 8], warpsPerCTA = [4, 1], order = [1, 0], CTAsPerCGA = [1, 1], CTASplitNum = [1, 1], CTAOrder = [0, 1]}> #blocked3 = #triton_gpu.blocked<{sizePerThread = [2, 2], threadsPerWarp = [4, 16], warpsPerCTA = [4, 1], order = [1, 0], CTAsPerCGA = [1, 1], CTASplitNum = [1, 1], CTAOrder = [1, 0]}> @@ -127,7 +129,8 @@ module attributes {"triton_gpu.compute-capability" = 0 : i32, "triton_gpu.num-ct } } -// CHECK: #mfma2 = #triton_gpu.mfma<{version = 2.0, warpsPerCTA = [1, 4], instrShape = [16, 16], isTransposed = false}> +// CHECK: #mfma2 +// CHECK-SAME: isTransposed = false #blocked4 = #triton_gpu.blocked<{sizePerThread = [1, 4], threadsPerWarp = [4, 16], warpsPerCTA = [4, 1], order = [1, 0], CTAsPerCGA = [1, 1], CTASplitNum = [1, 1], CTAOrder = [0, 1]}> #blocked5 = #triton_gpu.blocked<{sizePerThread = [1, 2], threadsPerWarp = [4, 16], warpsPerCTA = [4, 1], order = [1, 0], CTAsPerCGA = [1, 1], CTASplitNum = [1, 1], CTAOrder = [0, 1]}>