vllm/vllm/platforms
youkaichao eebad39f26
[torch.compile] support all attention backends (#10558)
Signed-off-by: youkaichao <youkaichao@gmail.com>
2024-11-22 14:04:42 -08:00
..
__init__.py [Platform][Refactor] Extract func get_default_attn_backend to Platform (#10358) 2024-11-19 11:22:26 +08:00
cpu.py [torch.compile] support all attention backends (#10558) 2024-11-22 14:04:42 -08:00
cuda.py [torch.compile] support all attention backends (#10558) 2024-11-22 14:04:42 -08:00
hpu.py [torch.compile] support all attention backends (#10558) 2024-11-22 14:04:42 -08:00
interface.py [torch.compile] support all attention backends (#10558) 2024-11-22 14:04:42 -08:00
neuron.py [platforms] absorb worker cls difference into platforms folder (#10555) 2024-11-21 21:00:32 -08:00
openvino.py [torch.compile] support all attention backends (#10558) 2024-11-22 14:04:42 -08:00
rocm.py [torch.compile] support all attention backends (#10558) 2024-11-22 14:04:42 -08:00
tpu.py [torch.compile] support all attention backends (#10558) 2024-11-22 14:04:42 -08:00
xpu.py [torch.compile] support all attention backends (#10558) 2024-11-22 14:04:42 -08:00