This website requires JavaScript.
Explore
Help
Register
Sign In
squall
/
vllm
Watch
1
Star
0
Fork
0
You've already forked vllm
Code
Issues
Pull Requests
Actions
1
Packages
Projects
Releases
Wiki
Activity
fb3db61688
vllm
/
vllm
/
attention
History
Thomas Parnell
9a7e2d0534
[Bugfix] Allow vllm to still work if triton is not installed. (
#6786
)
...
Signed-off-by: Thomas Parnell <tpa@zurich.ibm.com>
2024-07-29 14:51:27 -07:00
..
backends
[TPU] Reduce compilation time & Upgrade PyTorch XLA version (
#6856
)
2024-07-27 10:28:33 -07:00
ops
[Bugfix] Allow vllm to still work if triton is not installed. (
#6786
)
2024-07-29 14:51:27 -07:00
__init__.py
[Core] Refactor _prepare_model_input_tensors - take 2 (
#6164
)
2024-07-17 09:37:16 -07:00
layer.py
[Misc] Support FP8 kv cache scales from compressed-tensors (
#6528
)
2024-07-23 04:11:50 +00:00
selector.py
[Core] Refactor _prepare_model_input_tensors - take 2 (
#6164
)
2024-07-17 09:37:16 -07:00