vllm/tests/entrypoints/llm
Andy 2529d09b5a
[Frontend] Batch inference for llm.chat() API (#8648)
Co-authored-by: Cyrus Leung <cyrus.tl.leung@gmail.com>
Co-authored-by: Cyrus Leung <tlleungac@connect.ust.hk>
Co-authored-by: Roger Wang <ywang@roblox.com>
Co-authored-by: Roger Wang <136131678+ywang96@users.noreply.github.com>
2024-09-24 09:44:11 -07:00
..
__init__.py [CI/Build] [3/3] Reorganize entrypoints tests (#5966) 2024-06-30 12:58:49 +08:00
test_encode.py [CI/Build] [3/3] Reorganize entrypoints tests (#5966) 2024-06-30 12:58:49 +08:00
test_generate_multiple_loras.py [Core] Support load and unload LoRA in api server (#6566) 2024-09-05 18:10:33 -07:00
test_generate.py [Frontend] Batch inference for llm.chat() API (#8648) 2024-09-24 09:44:11 -07:00
test_guided_generate.py Support for guided decoding for offline LLM (#6878) 2024-08-04 03:12:09 +00:00
test_lazy_outlines.py [misc][core] lazy import outlines (#7831) 2024-08-24 00:51:38 -07:00
test_prompt_validation.py [BugFix] Fix server crash on empty prompt (#7746) 2024-08-23 13:12:44 +00:00