vllm/docs/source
2024-11-21 23:43:43 +00:00
..
_static [doc] fix location of runllm widget (#10266) 2024-11-12 14:34:39 -08:00
_templates/sections [Doc] Guide for adding multi-modal plugins (#6205) 2024-07-10 14:55:34 +08:00
assets [Doc] Add the start of an arch overview page (#10368) 2024-11-19 09:52:11 +00:00
automatic_prefix_caching [Doc] Add an automatic prefix caching section in vllm documentation (#5324) 2024-06-11 10:24:59 -07:00
community [Docs] Add Nebius as sponsors (#10371) 2024-11-15 12:47:40 -08:00
contributing [Doc] Move PR template content to docs (#10159) 2024-11-15 13:20:20 -08:00
design [Doc] Add the start of an arch overview page (#10368) 2024-11-19 09:52:11 +00:00
dev [Doc] Move CONTRIBUTING to docs site (#9924) 2024-11-08 05:15:12 +00:00
getting_started [Hardware][CPU] Support chunked-prefill and prefix-caching on CPU (#10355) 2024-11-20 10:57:39 +00:00
models Add openai.beta.chat.completions.parse example to structured_outputs.rst (#10433) 2024-11-19 04:37:46 +00:00
performance Doc: Improve benchmark documentation (#9927) 2024-11-06 23:50:35 -08:00
quantization [Hardware][XPU] AWQ/GPTQ support for xpu backend (#10107) 2024-11-18 11:18:05 -07:00
serving Add small example to metrics.rst (#10550) 2024-11-21 23:43:43 +00:00
conf.py Online video support for VLMs (#10020) 2024-11-07 20:25:59 +00:00
generate_examples.py Add example scripts to documentation (#4225) 2024-04-22 16:36:54 +00:00
index.rst [Doc] Add the start of an arch overview page (#10368) 2024-11-19 09:52:11 +00:00