For faster and smaller builds in some simple cases, provide switches to allow disabling -backward -alibi -uneven k -dropout -local attention Co-authored-by: Jeremy Francis Reizenstein <bottler@users.noreply.github.com> |
||
|---|---|---|
| .. | ||
| cutlass@751eb9a885 | ||
| flash_attn | ||
| ft_attention | ||
| fused_dense_lib | ||
| fused_softmax | ||
| layer_norm | ||
| rotary | ||
| xentropy | ||