This website requires JavaScript.
Explore
Help
Register
Sign In
squall
/
vllm
Watch
1
Star
0
Fork
0
You've already forked vllm
Code
Issues
Pull Requests
Actions
1
Packages
Projects
Releases
Wiki
Activity
aebfcb262a
vllm
/
vllm
/
model_executor
History
forpanyang
ab9e8488d5
Add Yi model to quantization support (
#1600
)
2023-11-09 11:47:14 -08:00
..
layers
Support YaRN models (
#1264
)
2023-11-03 14:12:48 -07:00
models
ChatGLM Support (
#1261
)
2023-11-06 16:09:33 -08:00
parallel_utils
Implement prompt logprobs & Batched topk for computing logprobs (
#1328
)
2023-10-16 10:56:50 -07:00
quantization_utils
Support SqueezeLLM (
#1326
)
2023-10-21 23:14:59 -07:00
__init__.py
[Quality] Add code formatter and linter (
#326
)
2023-07-03 11:31:55 -07:00
input_metadata.py
Delay GPU->CPU sync in sampling (
#1337
)
2023-10-30 09:01:34 -07:00
model_loader.py
Add Yi model to quantization support (
#1600
)
2023-11-09 11:47:14 -08:00
utils.py
TP/quantization/weight loading refactor part 1 - Simplify parallel linear logic (
#1181
)
2023-10-02 15:36:09 -07:00
weight_utils.py
Add blacklist in model checkpoint (
#1325
)
2023-10-12 01:05:37 -07:00