flash-attention/flash_attn/ops/triton
2023-04-18 21:51:35 -07:00
..
k_activations.py Add GPT and ViT models 2022-11-13 22:30:23 -08:00
linear.py Add GPT and ViT models 2022-11-13 22:30:23 -08:00
mlp.py Implement LLaMa 2023-04-18 21:51:35 -07:00