flash-attention/tests/models
dan_the_3rd 011ec323d6
Support MQA + MP for decoding (#490)
Co-authored-by: danthe3rd <danthe3rd>
2023-08-30 10:29:54 -07:00
..
test_baichuan.py FEAT: add codes which supporting for baichuan-inc/Baichuan-7B (#425) 2023-08-21 11:05:06 -07:00
test_bert.py [GPT] Fix loading weights from HF hub 2023-08-21 22:56:02 -07:00
test_falcon.py Run isort and black on test files 2023-08-18 20:59:35 -07:00
test_gpt_generation_parallel.py Run isort and black on test files 2023-08-18 20:59:35 -07:00
test_gpt_neox.py Run isort and black on test files 2023-08-18 20:59:35 -07:00
test_gpt_parallel.py Run isort and black on test files 2023-08-18 20:59:35 -07:00
test_gpt.py Support MQA + MP for decoding (#490) 2023-08-30 10:29:54 -07:00
test_gptj.py Run isort and black on test files 2023-08-18 20:59:35 -07:00
test_llama.py Run isort and black on test files 2023-08-18 20:59:35 -07:00
test_opt.py [GPT] Move GPT and OPT generation tests to test_{gpt,opt}.py 2023-08-26 12:55:02 -07:00
test_vit.py Run isort and black on test files 2023-08-18 20:59:35 -07:00