|
layers
|
Implement TensorParallel for MHA
|
2022-12-25 11:39:55 -08:00 |
|
models
|
Implement generation for GPT
|
2022-12-27 21:01:50 -08:00 |
|
modules
|
Implement generation for GPT
|
2022-12-27 21:01:50 -08:00 |
|
ops
|
Implement Tensor Parallel for transformer Block
|
2022-12-25 14:08:21 -08:00 |
|
utils
|
Implement generation for GPT
|
2022-12-27 21:01:50 -08:00 |
|
__init__.py
|
Add missing __init__.py
|
2022-07-03 02:04:55 -04:00 |
|
bert_padding.py
|
remove numpy dependency
|
2022-10-06 19:17:15 +02:00 |
|
flash_attention.py
|
Implement BERT
|
2022-12-18 21:47:27 -08:00 |
|
flash_blocksparse_attention.py
|
Rename src -> flash_attn
|
2022-06-01 18:50:26 -07:00 |