cutlass/examples/41_fused_multi_head_attention/gemm
dan_the_3rd f303889ed9
fMHA: Sync FW with xFormers (#828)
* fMHA: Add support for bias+dropout in FW

* Remove 'getMaximumSharedMemoryPerBlockKb'

* fix comments

---------

Co-authored-by: danthe3rd <danthe3rd>
Co-authored-by: Haicheng Wu <haichengw@nvidia.com>
2023-02-22 23:25:31 -05:00
..
custom_mma_base.h New updates for 2.11 (#775) 2023-01-20 16:32:57 -05:00
custom_mma_multistage.h New updates for 2.11 (#775) 2023-01-20 16:32:57 -05:00
custom_mma_pipelined.h New updates for 2.11 (#775) 2023-01-20 16:32:57 -05:00
custom_mma.h New updates for 2.11 (#775) 2023-01-20 16:32:57 -05:00
find_default_mma.h fMHA: Sync FW with xFormers (#828) 2023-02-22 23:25:31 -05:00
mma_accum_lambda_iterator.h fMHA: Sync FW with xFormers (#828) 2023-02-22 23:25:31 -05:00
mma_from_smem.h fMHA: Sync FW with xFormers (#828) 2023-02-22 23:25:31 -05:00