flash-attention/flash_attn
2023-08-21 22:56:02 -07:00
..
layers Run isort and black on python files 2023-08-18 14:22:11 -07:00
losses Run isort and black on python files 2023-08-18 14:22:11 -07:00
models FEAT: add codes which supporting for baichuan-inc/Baichuan-7B (#425) 2023-08-21 11:05:06 -07:00
modules Run isort and black on python files 2023-08-18 14:22:11 -07:00
ops Run isort and black on python files 2023-08-18 14:22:11 -07:00
utils [GPT] Fix loading weights from HF hub 2023-08-21 22:56:02 -07:00
__init__.py Run isort and black on python files 2023-08-18 14:22:11 -07:00
bert_padding.py Run isort and black on python files 2023-08-18 14:22:11 -07:00
flash_attn_interface.py Import torch before flash_attn_2_cuda 2023-08-19 21:07:33 -07:00
flash_attn_triton_og.py Run isort and black on python files 2023-08-18 14:22:11 -07:00
flash_attn_triton.py Run isort and black on python files 2023-08-18 14:22:11 -07:00
flash_blocksparse_attention.py Run isort and black on python files 2023-08-18 14:22:11 -07:00
flash_blocksparse_attn_interface.py Run isort and black on python files 2023-08-18 14:22:11 -07:00
fused_softmax.py Run isort and black on python files 2023-08-18 14:22:11 -07:00