vllm/benchmarks
2024-05-03 10:20:12 -07:00
..
kernels [Core][Model runner refactoring 1/N] Refactor attn metadata term (#4518) 2024-05-03 10:20:12 -07:00
backend_request_func.py [Bugfix] fix output parsing error for trtllm backend (#4137) 2024-04-17 11:07:23 +00:00
benchmark_latency.py [Bugfix] Get available quantization methods from quantization registry (#4098) 2024-04-18 00:21:55 -07:00
benchmark_prefix_caching.py [Core] Enable prefix caching with block manager v2 enabled (#4142) 2024-05-01 11:20:32 -07:00
benchmark_serving.py [Misc] Update ShareGPT Dataset Sampling in Serving Benchmark (#4279) 2024-04-24 09:49:13 -07:00
benchmark_throughput.py [Misc] Use public API in benchmark_throughput (#4300) 2024-04-24 21:10:24 +00:00
launch_tgi_server.sh Serving Benchmark Refactoring (#2433) 2024-02-12 22:53:00 -08:00
README.md Change the name to vLLM (#150) 2023-06-17 03:07:40 -07:00
sonnet.txt feat(benchmarks): Add Prefix Caching Benchmark to Serving Benchmark (#3277) 2024-03-27 13:39:26 -07:00

Benchmarking vLLM

Downloading the ShareGPT dataset

You can download the dataset by running:

wget https://huggingface.co/datasets/anon8231489123/ShareGPT_Vicuna_unfiltered/resolve/main/ShareGPT_V3_unfiltered_cleaned_split.json