vllm/vllm/engine
2024-09-24 05:45:20 +00:00
..
multiprocessing Revert "[Core] Rename PromptInputs to PromptType, and inputs to prompt" (#8750) 2024-09-24 05:45:20 +00:00
output_processor [Core] Optimize Async + Multi-step (#8050) 2024-09-03 18:50:29 +00:00
__init__.py Change the name to vLLM (#150) 2023-06-17 03:07:40 -07:00
arg_utils.py Add output streaming support to multi-step + async while ensuring RequestOutput obj reuse (#8335) 2024-09-23 15:38:04 -07:00
async_llm_engine.py Revert "[Core] Rename PromptInputs to PromptType, and inputs to prompt" (#8750) 2024-09-24 05:45:20 +00:00
async_timeout.py [Bugfix] AsyncLLMEngine hangs with asyncio.run (#5654) 2024-06-19 13:57:12 -07:00
llm_engine.py Revert "[Core] Rename PromptInputs to PromptType, and inputs to prompt" (#8750) 2024-09-24 05:45:20 +00:00
metrics_types.py [MISC] Add prefix cache hit rate to metrics (#7606) 2024-08-19 11:52:07 -07:00
metrics.py [MISC] Add prefix cache hit rate to metrics (#7606) 2024-08-19 11:52:07 -07:00
protocol.py Revert "[Core] Rename PromptInputs to PromptType, and inputs to prompt" (#8750) 2024-09-24 05:45:20 +00:00