vllm/docs/source/serving
youkaichao 2768884ac4
[Doc] Add note for docker user (#4340)
Co-authored-by: Simon Mo <simon.mo@hey.com>
2024-04-24 21:09:44 +00:00
..
deploying_with_bentoml.rst docs: Add BentoML deployment doc (#3336) 2024-03-12 10:34:30 -07:00
deploying_with_docker.rst [Doc] Add note for docker user (#4340) 2024-04-24 21:09:44 +00:00
deploying_with_kserve.rst docs: Add tutorial on deploying vLLM model with KServe (#2586) 2024-03-01 11:04:14 -08:00
deploying_with_triton.rst Add documentation to Triton server tutorial (#983) 2023-09-20 10:32:40 -07:00
distributed_serving.rst [Doc] Documentation for distributed inference (#261) 2023-06-26 11:34:23 -07:00
integrations.rst [Doc] Add docs about OpenAI compatible server (#3288) 2024-03-18 22:05:34 -07:00
metrics.rst Add Production Metrics in Prometheus format (#1890) 2023-12-02 16:37:44 -08:00
openai_compatible_server.md [Doc] Fix getting stared to use publicly available model (#3963) 2024-04-10 18:05:52 +00:00
run_on_sky.rst [Doc] Update the SkyPilot doc with serving and Llama-3 (#4276) 2024-04-22 15:34:31 -07:00
serving_with_langchain.rst docs: fix langchain (#2736) 2024-02-03 18:17:55 -08:00
usage_stats.md Usage Stats Collection (#2852) 2024-03-28 22:16:12 -07:00