cutlass/examples/41_fused_multi_head_attention/gemm
dan_the_3rd 146d314057
Update fMHA kernels (#992)
* Update fMHA kernels

Upstream recent changes to fMHA that we did in xFormers.
Previous version in CUTLASS: facebookresearch/xformers@b6be33a
Updating to: facebookresearch/xformers@55a4798

* minor changes

* make var work

---------

Co-authored-by: danthe3rd <danthe3rd>
Co-authored-by: Haicheng Wu <haichengw@nvidia.com>
2023-07-12 22:30:46 -04:00
..
custom_mma_base.h New updates for 2.11 (#775) 2023-01-20 16:32:57 -05:00
custom_mma_multistage.h Update fMHA kernels (#992) 2023-07-12 22:30:46 -04:00
custom_mma_pipelined.h Update fMHA kernels (#992) 2023-07-12 22:30:46 -04:00
custom_mma.h New updates for 2.11 (#775) 2023-01-20 16:32:57 -05:00
find_default_mma.h fMHA: Sync FW with xFormers (#828) 2023-02-22 23:25:31 -05:00
mma_accum_lambda_iterator.h fMHA: Sync FW with xFormers (#828) 2023-02-22 23:25:31 -05:00
mma_from_smem.h Update fMHA kernels (#992) 2023-07-12 22:30:46 -04:00