flash-attention/csrc/flash_attn
Antoni Viros 83e41b3ca4
Add custom ops for compatibility with PT Compile (#1139)
* Add custom ops for compatibility with PT Compile

* Add support for varlen functions too

* Add version checks for pytorch API

* Fix PT compile interfaces so it works e2e

* Make sure PT < 2.4 runs fine

* Fix python mistake

* Fix all the autograd magic issues

* typo on head_dim

* Fix deterministic test failures, remove unneeded detaches()

* remove test requires_grad

* Resolve all the pytorch versioning issues

* C++ and python refactor to improve padding management for torch.compile()

* Add improvements suggested by @anijain2305
2024-09-17 19:49:26 -07:00
..
src Split bwd into more .cu files to speed up compilation 2024-07-23 01:32:09 -07:00
flash_api.cpp Add custom ops for compatibility with PT Compile (#1139) 2024-09-17 19:49:26 -07:00