vllm/csrc
Kevin H. Luu aaccca2b4d
[CI/Build] Fix machete generated kernel files ordering (#8976)
Signed-off-by: kevin <kevin@anyscale.com>
Co-authored-by: Cody Yu <hao.yu.cody@gmail.com>
2024-10-01 03:33:12 +00:00
..
attention [AMD][CI/Build] Disambiguation of the function call for ROCm 6.2 headers compatibility (#7477) 2024-08-21 16:47:36 -07:00
core [Bugfix] Fix Marlin MoE act order when is_k_full == False (#8741) 2024-09-28 18:19:40 -07:00
cpu [Kernel] AQ AZP 3/4: Asymmetric quantization kernels (#7270) 2024-09-16 11:52:40 -07:00
cutlass_extensions [Kernel] (2/N) Machete - Integrate into CompressedTensorsWNA16 and GPTQMarlin (#7701) 2024-09-23 13:46:26 -04:00
mamba [Kernel][Model] Varlen prefill + Prefill chunking support for mamba kernels and Jamba model (#8533) 2024-09-29 17:35:58 -04:00
moe [Bugfix] Fix Marlin MoE act order when is_k_full == False (#8741) 2024-09-28 18:19:40 -07:00
prepare_inputs [Core] Multi-Step + Single Step Prefills via Chunked Prefill code path (#8378) 2024-09-27 13:32:07 -07:00
quantization [CI/Build] Fix machete generated kernel files ordering (#8976) 2024-10-01 03:33:12 +00:00
rocm [Kernel][Amd] Add fp8 kv cache support for rocm custom paged attention (#8577) 2024-09-19 17:37:57 +00:00
activation_kernels.cu [Model] Port over CLIPVisionModel for VLMs (#5591) 2024-06-20 11:52:09 +00:00
cache_kernels.cu Add fp8 support to reshape_and_cache_flash (#6667) 2024-07-24 18:36:52 +00:00
cache.h Add fp8 support to reshape_and_cache_flash (#6667) 2024-07-24 18:36:52 +00:00
cuda_compat.h [Kernel][ROCm][AMD] enable fused topk_softmax kernel for moe layer (#4927) 2024-06-02 14:13:26 -07:00
cuda_utils_kernels.cu [Kernel][Misc] Use TORCH_LIBRARY instead of PYBIND11_MODULE for custom ops (#5047) 2024-06-09 16:23:30 -04:00
cuda_utils.h [Kernel] (1/N) Machete - Hopper Optimized Mixed Precision Linear Kernel (#7174) 2024-08-20 07:09:33 -06:00
custom_all_reduce_test.cu [Bugfix][Kernel] Implement acquire/release polyfill for Pascal (#8776) 2024-09-24 21:26:33 -07:00
custom_all_reduce.cu [torch.compile] register allreduce operations as custom ops (#8526) 2024-09-16 22:57:57 -07:00
custom_all_reduce.cuh [Bugfix][Kernel] Implement acquire/release polyfill for Pascal (#8776) 2024-09-24 21:26:33 -07:00
dispatch_utils.h [Kernel][Misc] Use TORCH_LIBRARY instead of PYBIND11_MODULE for custom ops (#5047) 2024-06-09 16:23:30 -04:00
layernorm_kernels.cu [Kernel] Replaced blockReduce[...] functions with cub::BlockReduce (#7233) 2024-08-21 20:18:00 -04:00
moe_align_block_size_kernels.cu [Kernel][Misc] Use TORCH_LIBRARY instead of PYBIND11_MODULE for custom ops (#5047) 2024-06-09 16:23:30 -04:00
ops.h [Kernel][Model] Varlen prefill + Prefill chunking support for mamba kernels and Jamba model (#8533) 2024-09-29 17:35:58 -04:00
permute_cols.cu [Kernel] (2/N) Machete - Integrate into CompressedTensorsWNA16 and GPTQMarlin (#7701) 2024-09-23 13:46:26 -04:00
pos_encoding_kernels.cu [Kernel][Misc] Use TORCH_LIBRARY instead of PYBIND11_MODULE for custom ops (#5047) 2024-06-09 16:23:30 -04:00
torch_bindings.cpp [Kernel][Model] Varlen prefill + Prefill chunking support for mamba kernels and Jamba model (#8533) 2024-09-29 17:35:58 -04:00