vllm/vllm
2023-11-01 02:14:18 -07:00
..
core Fix type hints (#1427) 2023-10-20 08:50:47 -07:00
engine [BugFix] Set engine_use_ray=True when TP>1 (#1531) 2023-11-01 02:14:18 -07:00
entrypoints Add support for spaces_between_special_tokens 2023-10-30 16:52:56 -07:00
model_executor Add MptForCausalLM key in model_loader (#1526) 2023-10-31 15:46:53 -07:00
transformers_utils Add support for spaces_between_special_tokens 2023-10-30 16:52:56 -07:00
worker Fix integer overflows in attention & cache ops (#1514) 2023-10-31 15:19:30 -07:00
__init__.py Bump up the version to v0.2.1 (#1355) 2023-10-16 12:58:57 -07:00
block.py [Quality] Add code formatter and linter (#326) 2023-07-03 11:31:55 -07:00
config.py Support SqueezeLLM (#1326) 2023-10-21 23:14:59 -07:00
logger.py [Fix] Fix duplicated logging messages (#1524) 2023-10-31 09:04:47 -07:00
outputs.py Implement prompt logprobs & Batched topk for computing logprobs (#1328) 2023-10-16 10:56:50 -07:00
py.typed Add py.typed so consumers of vLLM can get type checking (#1509) 2023-10-30 14:50:47 -07:00
sampling_params.py Add support for spaces_between_special_tokens 2023-10-30 16:52:56 -07:00
sequence.py [BugFix] Define __eq__ in SequenceGroupOutputs (#1389) 2023-10-17 01:09:44 -07:00
utils.py Allocate more shared memory to attention kernel (#1154) 2023-09-26 22:27:13 -07:00