|
block.py
|
[GPT] Enable FlashAttention for GPT-J
|
2023-07-21 17:29:10 -07:00 |
|
embedding.py
|
Reorder LN in Block, support OPT
|
2023-01-15 22:14:31 -08:00 |
|
mha.py
|
FlashAttention-2 release
|
2023-07-17 06:21:34 -07:00 |
|
mlp.py
|
Implement LLaMa
|
2023-04-18 21:51:35 -07:00 |