vllm/docs/source/serving
2024-06-27 10:07:07 -07:00
..
deploying_with_bentoml.rst docs: Add BentoML deployment doc (#3336) 2024-03-12 10:34:30 -07:00
deploying_with_cerebrium.rst [Doc] Added cerebrium as Integration option (#5553) 2024-06-18 15:56:59 -07:00
deploying_with_docker.rst [Doc] Update docker references (#5614) 2024-06-19 15:01:45 -07:00
deploying_with_dstack.rst add doc about serving option on dstack (#3074) 2024-05-30 10:11:07 -07:00
deploying_with_kserve.rst docs: Add tutorial on deploying vLLM model with KServe (#2586) 2024-03-01 11:04:14 -08:00
deploying_with_lws.rst Support to serve vLLM on Kubernetes with LWS (#4829) 2024-05-16 16:37:29 -07:00
deploying_with_triton.rst Add documentation to Triton server tutorial (#983) 2023-09-20 10:32:40 -07:00
distributed_serving.rst [doc][faq] add warning to download models for every nodes (#5783) 2024-06-24 15:37:42 +08:00
env_vars.rst [doc][misc] add note for Kubernetes users (#5916) 2024-06-27 10:07:07 -07:00
integrations.rst [Doc] Added cerebrium as Integration option (#5553) 2024-06-18 15:56:59 -07:00
metrics.rst Add Production Metrics in Prometheus format (#1890) 2023-12-02 16:37:44 -08:00
openai_compatible_server.md [Frontend] Add OpenAI Vision API Support (#5237) 2024-06-07 11:23:32 -07:00
run_on_sky.rst [Doc] Update the SkyPilot doc with serving and Llama-3 (#4276) 2024-04-22 15:34:31 -07:00
serving_with_langchain.rst docs: fix langchain (#2736) 2024-02-03 18:17:55 -08:00
tensorizer.rst [Doc] Update documentation on Tensorizer (#5471) 2024-06-14 11:27:57 -07:00
usage_stats.md Usage Stats Collection (#2852) 2024-03-28 22:16:12 -07:00