This website requires JavaScript.
Explore
Help
Register
Sign In
squall
/
vllm
Watch
1
Star
0
Fork
0
You've already forked vllm
Code
Issues
Pull Requests
Actions
1
Packages
Projects
Releases
Wiki
Activity
4026a049d3
vllm
/
vllm
/
engine
History
Zhuohan Li
0b7db411b5
[Bug] Fix the OOM condition for CPU cache (
#260
)
2023-06-26 11:16:13 -07:00
..
__init__.py
Change the name to vLLM (
#150
)
2023-06-17 03:07:40 -07:00
arg_utils.py
Reduce GPU memory utilization to make sure OOM doesn't happen (
#153
)
2023-06-18 17:33:50 +08:00
async_llm_engine.py
[Bugfix] Fix a bug in RequestOutput.finished (
#202
)
2023-06-22 00:17:24 -07:00
llm_engine.py
[Bug] Fix the OOM condition for CPU cache (
#260
)
2023-06-26 11:16:13 -07:00
ray_utils.py
Change the name to vLLM (
#150
)
2023-06-17 03:07:40 -07:00
tokenizer_utils.py
fix wrong using getattr to get dict value (
#232
)
2023-06-24 22:00:24 -07:00