This website requires JavaScript.
Explore
Help
Register
Sign In
squall
/
vllm
Watch
1
Star
0
Fork
0
You've already forked vllm
Code
Issues
Pull Requests
Actions
1
Packages
Projects
Releases
Wiki
Activity
02a43f82a9
vllm
/
vllm
/
platforms
History
youkaichao
eebad39f26
[torch.compile] support all attention backends (
#10558
)
...
Signed-off-by: youkaichao <youkaichao@gmail.com>
2024-11-22 14:04:42 -08:00
..
__init__.py
[Platform][Refactor] Extract func
get_default_attn_backend
to
Platform
(
#10358
)
2024-11-19 11:22:26 +08:00
cpu.py
[torch.compile] support all attention backends (
#10558
)
2024-11-22 14:04:42 -08:00
cuda.py
[torch.compile] support all attention backends (
#10558
)
2024-11-22 14:04:42 -08:00
hpu.py
[torch.compile] support all attention backends (
#10558
)
2024-11-22 14:04:42 -08:00
interface.py
[torch.compile] support all attention backends (
#10558
)
2024-11-22 14:04:42 -08:00
neuron.py
[platforms] absorb worker cls difference into platforms folder (
#10555
)
2024-11-21 21:00:32 -08:00
openvino.py
[torch.compile] support all attention backends (
#10558
)
2024-11-22 14:04:42 -08:00
rocm.py
[torch.compile] support all attention backends (
#10558
)
2024-11-22 14:04:42 -08:00
tpu.py
[torch.compile] support all attention backends (
#10558
)
2024-11-22 14:04:42 -08:00
xpu.py
[torch.compile] support all attention backends (
#10558
)
2024-11-22 14:04:42 -08:00