vllm/vllm/engine
2023-06-26 11:16:13 -07:00
..
__init__.py Change the name to vLLM (#150) 2023-06-17 03:07:40 -07:00
arg_utils.py Reduce GPU memory utilization to make sure OOM doesn't happen (#153) 2023-06-18 17:33:50 +08:00
async_llm_engine.py [Bugfix] Fix a bug in RequestOutput.finished (#202) 2023-06-22 00:17:24 -07:00
llm_engine.py [Bug] Fix the OOM condition for CPU cache (#260) 2023-06-26 11:16:13 -07:00
ray_utils.py Change the name to vLLM (#150) 2023-06-17 03:07:40 -07:00
tokenizer_utils.py fix wrong using getattr to get dict value (#232) 2023-06-24 22:00:24 -07:00