| .. |
|
_static
|
[doc] fix location of runllm widget (#10266)
|
2024-11-12 14:34:39 -08:00 |
|
_templates/sections
|
[Doc] Guide for adding multi-modal plugins (#6205)
|
2024-07-10 14:55:34 +08:00 |
|
assets
|
[Doc] Add the start of an arch overview page (#10368)
|
2024-11-19 09:52:11 +00:00 |
|
automatic_prefix_caching
|
[Doc] Add an automatic prefix caching section in vllm documentation (#5324)
|
2024-06-11 10:24:59 -07:00 |
|
community
|
[Docs] Add Nebius as sponsors (#10371)
|
2024-11-15 12:47:40 -08:00 |
|
contributing
|
[Doc] Move PR template content to docs (#10159)
|
2024-11-15 13:20:20 -08:00 |
|
design
|
[Doc] Add the start of an arch overview page (#10368)
|
2024-11-19 09:52:11 +00:00 |
|
dev
|
[Doc] Move CONTRIBUTING to docs site (#9924)
|
2024-11-08 05:15:12 +00:00 |
|
getting_started
|
[Hardware][CPU] Support chunked-prefill and prefix-caching on CPU (#10355)
|
2024-11-20 10:57:39 +00:00 |
|
models
|
Add openai.beta.chat.completions.parse example to structured_outputs.rst (#10433)
|
2024-11-19 04:37:46 +00:00 |
|
performance
|
Doc: Improve benchmark documentation (#9927)
|
2024-11-06 23:50:35 -08:00 |
|
quantization
|
[Hardware][XPU] AWQ/GPTQ support for xpu backend (#10107)
|
2024-11-18 11:18:05 -07:00 |
|
serving
|
Add small example to metrics.rst (#10550)
|
2024-11-21 23:43:43 +00:00 |
|
conf.py
|
Online video support for VLMs (#10020)
|
2024-11-07 20:25:59 +00:00 |
|
generate_examples.py
|
Add example scripts to documentation (#4225)
|
2024-04-22 16:36:54 +00:00 |
|
index.rst
|
[Doc] Add the start of an arch overview page (#10368)
|
2024-11-19 09:52:11 +00:00 |