According to the `setup.py` file, only dependencies are torch and einops. But the `bert_padding.py` file requires `numpy` only to multiply the elements of a `torch.Size` object. This change aims at allowing the use of FlashAttention without numpy. |
||
|---|---|---|
| .. | ||
| __init__.py | ||
| bert_padding.py | ||
| flash_attention.py | ||
| flash_attn_interface.py | ||
| flash_blocksparse_attention.py | ||
| flash_blocksparse_attn_interface.py | ||
| rotary.py | ||