vllm/vllm/entrypoints
Robert Shaw ed812a73fa
[ Frontend ] Multiprocessing for OpenAI Server with zeromq (#6883)
Signed-off-by: Joe Runde <Joseph.Runde@ibm.com>
Co-authored-by: Joe Runde <Joseph.Runde@ibm.com>
Co-authored-by: Joe Runde <joe@joerun.de>
Co-authored-by: Nick Hill <nickhill@us.ibm.com>
Co-authored-by: Simon Mo <simon.mo@hey.com>
2024-08-02 18:27:28 -07:00
..
openai [ Frontend ] Multiprocessing for OpenAI Server with zeromq (#6883) 2024-08-02 18:27:28 -07:00
__init__.py Change the name to vLLM (#150) 2023-06-17 03:07:40 -07:00
api_server.py Revert "[Frontend] Factor out code for running uvicorn" (#7012) 2024-07-31 16:34:26 -07:00
chat_utils.py [Model] Initialize support for InternVL2 series models (#6514) 2024-07-29 10:16:30 +00:00
llm.py [Frontend] Refactor prompt processing (#4028) 2024-07-22 10:13:53 -07:00
logger.py [Frontend] Refactor prompt processing (#4028) 2024-07-22 10:13:53 -07:00