vllm/vllm
2023-11-09 11:47:14 -08:00
..
core Fix type hints (#1427) 2023-10-20 08:50:47 -07:00
engine [BugFix] Set engine_use_ray=True when TP>1 (#1531) 2023-11-01 02:14:18 -07:00
entrypoints Add /health Endpoint for both Servers (#1540) 2023-11-01 10:29:44 -07:00
model_executor Add Yi model to quantization support (#1600) 2023-11-09 11:47:14 -08:00
transformers_utils ChatGLM Support (#1261) 2023-11-06 16:09:33 -08:00
worker Fix input_metadata.selected_token_indices in worker prepare_inputs (#1546) 2023-11-08 14:19:12 -08:00
__init__.py Bump up the version to v0.2.1 (#1355) 2023-10-16 12:58:57 -07:00
block.py [Quality] Add code formatter and linter (#326) 2023-07-03 11:31:55 -07:00
config.py ChatGLM Support (#1261) 2023-11-06 16:09:33 -08:00
logger.py [Fix] Fix duplicated logging messages (#1524) 2023-10-31 09:04:47 -07:00
outputs.py docs: add description (#1553) 2023-11-03 09:14:52 -07:00
py.typed Add py.typed so consumers of vLLM can get type checking (#1509) 2023-10-30 14:50:47 -07:00
sampling_params.py Added logits processor API to sampling params (#1469) 2023-11-03 14:12:15 -07:00
sequence.py [BugFix] Define __eq__ in SequenceGroupOutputs (#1389) 2023-10-17 01:09:44 -07:00
utils.py Allocate more shared memory to attention kernel (#1154) 2023-09-26 22:27:13 -07:00