diff --git a/samples/99_flashattention/notes.txt b/samples/99_flashattention/notes.txt new file mode 100644 index 0000000..85e38cc --- /dev/null +++ b/samples/99_flashattention/notes.txt @@ -0,0 +1,5 @@ +https://github.com/intel/mlir-extensions/blob/main/test/Integration/Dialect/XeGPU/flash_attention_fwd.mlir + +https://github.com/intel/intel-extension-for-pytorch/blob/xpu-main/csrc/gpu/aten/operators/xetla/kernels/SDP/fmha_forward.hpp + +https://github.com/NVIDIA/cudnn-frontend/blob/main/docs/operations/Attention.md