flash-attention/tests/models
2023-09-16 01:20:16 -07:00
..
test_baichuan.py [Gen] Fix calling update_graph_cache in tests 2023-09-10 17:22:37 -07:00
test_bert.py Add BigCode converters (#532) 2023-09-10 17:24:50 -07:00
test_bigcode.py Add BigCode converters (#532) 2023-09-10 17:24:50 -07:00
test_falcon.py [Gen] Fix calling update_graph_cache in tests 2023-09-10 17:22:37 -07:00
test_gpt_generation_parallel.py [Gen] Refactor decoding function 2023-09-04 17:01:38 -07:00
test_gpt_neox.py Add tests for Pythia, GPT-JT, and RedPajama models 2023-09-13 01:10:39 -07:00
test_gpt_parallel.py Run isort and black on test files 2023-08-18 20:59:35 -07:00
test_gpt.py Implement rotary embedding in flash_attn_with_kvcache 2023-09-16 01:20:16 -07:00
test_gptj.py Add tests for Pythia, GPT-JT, and RedPajama models 2023-09-13 01:10:39 -07:00
test_llama.py [Gen] Fix calling update_graph_cache in tests 2023-09-10 17:22:37 -07:00
test_opt.py [Gen] Fix calling update_graph_cache in tests 2023-09-10 17:22:37 -07:00
test_vit.py Run isort and black on test files 2023-08-18 20:59:35 -07:00