vllm/vllm/entrypoints
Nick Hill 4289cad37f
[Frontend] Minor optimizations to zmq decoupled front-end (#7957)
Co-authored-by: Robert Shaw <rshaw@neuralmagic>
2024-08-28 17:22:43 -07:00
..
openai [Frontend] Minor optimizations to zmq decoupled front-end (#7957) 2024-08-28 17:22:43 -07:00
__init__.py Change the name to vLLM (#150) 2023-06-17 03:07:40 -07:00
api_server.py [mypy] Enable following imports for entrypoints (#7248) 2024-08-20 23:28:21 -07:00
chat_utils.py [Model] Add Mistral Tokenization to improve robustness and chat encoding (#7739) 2024-08-27 12:40:02 +00:00
launcher.py [Bugfix][Frontend] Fix Issues Under High Load With zeromq Frontend (#7394) 2024-08-21 13:34:14 -04:00
llm.py [Model] Add Mistral Tokenization to improve robustness and chat encoding (#7739) 2024-08-27 12:40:02 +00:00
logger.py [Frontend] Refactor prompt processing (#4028) 2024-07-22 10:13:53 -07:00