vllm/docs/source
2024-05-31 17:20:19 -07:00
..
assets [Doc] add visualization for multi-stage dockerfile (#4456) 2024-04-30 17:41:59 +00:00
community [Docs] Add Dropbox as sponsors (#5089) 2024-05-28 10:29:09 -07:00
dev [Core] Consolidate prompt arguments to LLM engines (#4328) 2024-05-28 13:29:31 -07:00
getting_started [Doc] add ccache guide in doc (#5012) 2024-05-23 23:21:54 +00:00
models [Doc] Add checkmark for GPTBigCodeForCausalLM LoRA support (#5171) 2024-05-31 17:20:19 -07:00
quantization Enable scaled FP8 (e4m3fn) KV cache on ROCm (AMD GPU) (#3290) 2024-04-03 14:15:55 -07:00
serving add doc about serving option on dstack (#3074) 2024-05-30 10:11:07 -07:00
conf.py [Doc] Use intersphinx and update entrypoints docs (#5125) 2024-05-30 09:59:23 -07:00
generate_examples.py Add example scripts to documentation (#4225) 2024-04-22 16:36:54 +00:00
index.rst [Core] Consolidate prompt arguments to LLM engines (#4328) 2024-05-28 13:29:31 -07:00