Skip to content

Commit

Permalink
refine chain-dot.mlir
Browse files Browse the repository at this point in the history
  • Loading branch information
scxiao committed Jan 25, 2024
1 parent cb69282 commit 0d2a73e
Showing 1 changed file with 6 additions and 3 deletions.
9 changes: 6 additions & 3 deletions test/TritonGPU/chain-dot.mlir
Original file line number Diff line number Diff line change
@@ -1,6 +1,7 @@
// RUN: triton-opt %s -split-input-file --tritonamdgpu-accelerate-matmul=arch-generation-name=gfx90a | FileCheck --check-prefixes=CHECK %s

// CHECK: #mfma = #triton_gpu.mfma<{version = 2.0, warpsPerCTA = [4, 1], instrShape = [32, 32], isTransposed = true}>
// CHECK: #mfma
// CHECK-SAME: isTransposed = true

#blocked = #triton_gpu.blocked<{sizePerThread = [1, 4], threadsPerWarp = [8, 8], warpsPerCTA = [4, 1], order = [1, 0], CTAsPerCGA = [1, 1], CTASplitNum = [1, 1], CTAOrder = [0, 1]}>
#blocked1 = #triton_gpu.blocked<{sizePerThread = [2, 2], threadsPerWarp = [4, 16], warpsPerCTA = [4, 1], order = [1, 0], CTAsPerCGA = [1, 1], CTASplitNum = [1, 1], CTAOrder = [1, 0]}>
Expand Down Expand Up @@ -64,7 +65,8 @@ module attributes {"triton_gpu.compute-capability" = 0 : i32, "triton_gpu.num-ct
}


// CHECK: #mfma1 = #triton_gpu.mfma<{version = 2.0, warpsPerCTA = [4, 1], instrShape = [32, 32], isTransposed = false}>
// CHECK: #mfma1
// CHECK-SAME: isTransposed = false

#blocked2 = #triton_gpu.blocked<{sizePerThread = [1, 4], threadsPerWarp = [8, 8], warpsPerCTA = [4, 1], order = [1, 0], CTAsPerCGA = [1, 1], CTASplitNum = [1, 1], CTAOrder = [0, 1]}>
#blocked3 = #triton_gpu.blocked<{sizePerThread = [2, 2], threadsPerWarp = [4, 16], warpsPerCTA = [4, 1], order = [1, 0], CTAsPerCGA = [1, 1], CTASplitNum = [1, 1], CTAOrder = [1, 0]}>
Expand Down Expand Up @@ -127,7 +129,8 @@ module attributes {"triton_gpu.compute-capability" = 0 : i32, "triton_gpu.num-ct
}
}

// CHECK: #mfma2 = #triton_gpu.mfma<{version = 2.0, warpsPerCTA = [1, 4], instrShape = [16, 16], isTransposed = false}>
// CHECK: #mfma2
// CHECK-SAME: isTransposed = false

#blocked4 = #triton_gpu.blocked<{sizePerThread = [1, 4], threadsPerWarp = [4, 16], warpsPerCTA = [4, 1], order = [1, 0], CTAsPerCGA = [1, 1], CTASplitNum = [1, 1], CTAOrder = [0, 1]}>
#blocked5 = #triton_gpu.blocked<{sizePerThread = [1, 2], threadsPerWarp = [4, 16], warpsPerCTA = [4, 1], order = [1, 0], CTAsPerCGA = [1, 1], CTASplitNum = [1, 1], CTAOrder = [0, 1]}>
Expand Down

0 comments on commit 0d2a73e

Please sign in to comment.