2023-06-17 18:07:40 +08:00
|
|
|
Welcome to vLLM!
|
|
|
|
|
================
|
2023-05-23 08:02:44 +08:00
|
|
|
|
2023-06-18 01:26:12 +08:00
|
|
|
vLLM is a high-throughput and memory-efficient inference and serving engine for large language models (LLM).
|
|
|
|
|
|
2023-05-23 08:02:44 +08:00
|
|
|
Documentation
|
|
|
|
|
-------------
|
|
|
|
|
|
|
|
|
|
.. toctree::
|
|
|
|
|
:maxdepth: 1
|
|
|
|
|
:caption: Getting Started
|
|
|
|
|
|
|
|
|
|
getting_started/installation
|
|
|
|
|
getting_started/quickstart
|
2023-06-03 13:35:17 +08:00
|
|
|
|
|
|
|
|
.. toctree::
|
|
|
|
|
:maxdepth: 1
|
|
|
|
|
:caption: Models
|
|
|
|
|
|
|
|
|
|
models/supported_models
|
|
|
|
|
models/adding_model
|