vllm/tests
Zhuohan Li db09d4ad83
[FIX] Fix Alibi implementation in PagedAttention kernel (#945)
* [FIX] Fix Alibi implementation in PagedAttention kernel

* Fix test_attention

* Fix

---------

Co-authored-by: Woosuk Kwon <woosuk.kwon@berkeley.edu>
Co-authored-by: Oliver-ss <yuansongwx@outlook.com>
2023-09-07 15:53:14 -07:00
..
async_engine Make AsyncLLMEngine more robust & fix batched abort (#969) 2023-09-07 13:43:45 -07:00
kernels [FIX] Fix Alibi implementation in PagedAttention kernel (#945) 2023-09-07 15:53:14 -07:00
models Add tests for models (#922) 2023-09-01 11:19:43 +09:00
samplers Align vLLM's beam search implementation with HF generate (#857) 2023-09-04 17:29:42 -07:00
conftest.py Use queue for finished requests (#957) 2023-09-05 19:27:23 -07:00