vllm/vllm
2023-06-29 22:14:17 -07:00
..
core [BugFix] Fix a bug in counting running sequences (#266) 2023-06-26 13:09:02 -07:00
engine [Tokenizer] Add tokenizer mode (#298) 2023-06-28 14:19:22 -07:00
entrypoints [Tokenizer] Add tokenizer mode (#298) 2023-06-28 14:19:22 -07:00
model_executor [Fix] Weight loading for GPTBigCode (#313) 2023-06-29 22:14:17 -07:00
transformers_utils [Tokenizer] Add tokenizer mode (#298) 2023-06-28 14:19:22 -07:00
worker [Fix] Do not pin memory when in WSL (#312) 2023-06-29 15:00:21 -07:00
__init__.py Bump up version to 0.1.1 (#204) 2023-06-22 15:33:32 +08:00
block.py Change the name to vLLM (#150) 2023-06-17 03:07:40 -07:00
config.py [Tokenizer] Add tokenizer mode (#298) 2023-06-28 14:19:22 -07:00
logger.py Change the name to vLLM (#150) 2023-06-17 03:07:40 -07:00
outputs.py [Fix] Better error message when there is OOM during cache initialization (#203) 2023-06-22 15:30:06 +08:00
sampling_params.py remove floats == 0 comparison (#285) 2023-06-28 14:11:51 -07:00
sequence.py Change the name to vLLM (#150) 2023-06-17 03:07:40 -07:00
utils.py [Fix] Do not pin memory when in WSL (#312) 2023-06-29 15:00:21 -07:00