vllm/benchmarks/kernels
2024-03-13 13:45:26 -07:00
..
benchmark_mixtral_moe.py Optimize Triton MoE Kernel (#2979) 2024-02-26 13:48:56 -08:00
benchmark_paged_attention.py Remove hardcoded device="cuda" to support more devices (#2503) 2024-02-01 15:46:39 -08:00
benchmark_rope.py Add batched RoPE kernel (#3095) 2024-03-13 13:45:26 -07:00