vllm/vllm
2023-11-27 11:18:26 -08:00
..
core [Fix] Fix bugs in scheduler (#1727) 2023-11-20 16:10:50 -08:00
engine [DOCS] Add engine args documentation (#1741) 2023-11-22 12:31:27 -08:00
entrypoints Added echo function to OpenAI API server. (#1504) 2023-11-26 21:29:17 -08:00
model_executor Init model on GPU to reduce CPU memory footprint (#1796) 2023-11-27 11:18:26 -08:00
transformers_utils Migrate linter from pylint to ruff (#1665) 2023-11-20 11:58:01 -08:00
worker [Build] Avoid building too many extensions (#1624) 2023-11-23 16:31:19 -08:00
__init__.py Bump up to v0.2.2 (#1689) 2023-11-18 21:57:07 -08:00
block.py [Quality] Add code formatter and linter (#326) 2023-07-03 11:31:55 -07:00
config.py fix RAM OOM when load large models in tensor parallel mode. (#1395) 2023-11-20 19:02:42 -08:00
logger.py [Fix] Fix duplicated logging messages (#1524) 2023-10-31 09:04:47 -07:00
outputs.py docs: add description (#1553) 2023-11-03 09:14:52 -07:00
py.typed Add py.typed so consumers of vLLM can get type checking (#1509) 2023-10-30 14:50:47 -07:00
sampling_params.py Fix repetition penalty aligned with huggingface (#1577) 2023-11-22 14:41:44 -08:00
sequence.py [BugFix] Define __eq__ in SequenceGroupOutputs (#1389) 2023-10-17 01:09:44 -07:00
utils.py [Build] Avoid building too many extensions (#1624) 2023-11-23 16:31:19 -08:00