vllm/tests
Antoni Baum c07ece5ca4
Make AsyncLLMEngine more robust & fix batched abort (#969)
Signed-off-by: Antoni Baum <antoni.baum@protonmail.com>
Co-authored-by: Avnish Narayan <38871737+avnishn@users.noreply.github.com>
2023-09-07 13:43:45 -07:00
..
async_engine Make AsyncLLMEngine more robust & fix batched abort (#969) 2023-09-07 13:43:45 -07:00
kernels [BugFix] Implement RoPE for GPT-J (#941) 2023-09-06 11:54:33 +09:00
models Add tests for models (#922) 2023-09-01 11:19:43 +09:00
samplers Align vLLM's beam search implementation with HF generate (#857) 2023-09-04 17:29:42 -07:00
conftest.py Use queue for finished requests (#957) 2023-09-05 19:27:23 -07:00