vllm/csrc/attention
Aleksandr Malyshev 812c981fa0
Splitting attention kernel file (#10091)
Signed-off-by: maleksan85 <maleksan@amd.com>
Co-authored-by: Aleksandr Malyshev <maleksan@amd.com>
2024-11-11 22:55:07 -08:00
..
attention_dtypes.h Enable scaled FP8 (e4m3fn) KV cache on ROCm (AMD GPU) (#3290) 2024-04-03 14:15:55 -07:00
attention_generic.cuh [CI/Build] Enforce style for C++ and CUDA code with clang-format (#4722) 2024-05-22 07:18:41 +00:00
attention_kernels.cuh Splitting attention kernel file (#10091) 2024-11-11 22:55:07 -08:00
attention_utils.cuh [AMD][CI/Build] Disambiguation of the function call for ROCm 6.2 headers compatibility (#7477) 2024-08-21 16:47:36 -07:00
dtype_bfloat16.cuh [CI/Build] Suppress divide-by-zero and missing return statement warnings (#7001) 2024-08-05 16:00:01 -04:00
dtype_float16.cuh [CI/Build] Enforce style for C++ and CUDA code with clang-format (#4722) 2024-05-22 07:18:41 +00:00
dtype_float32.cuh [CI/Build] Enforce style for C++ and CUDA code with clang-format (#4722) 2024-05-22 07:18:41 +00:00
dtype_fp8.cuh [CI/Build] Enforce style for C++ and CUDA code with clang-format (#4722) 2024-05-22 07:18:41 +00:00
paged_attention_v1.cu Splitting attention kernel file (#10091) 2024-11-11 22:55:07 -08:00
paged_attention_v2.cu Splitting attention kernel file (#10091) 2024-11-11 22:55:07 -08:00