vllm/vllm
2024-09-25 14:46:22 -07:00
..
adapter_commons [CI/Build] Update Ruff version (#8469) 2024-09-18 11:00:56 +00:00
assets [Model][VLM] Add LLaVA-Onevision model support (#8486) 2024-09-22 10:51:44 -07:00
attention [Kernel] Build flash-attn from source (#8245) 2024-09-20 23:27:10 -07:00
compilation [torch.compile] fix functionalization (#8480) 2024-09-14 09:46:04 -07:00
core [Misc] Fix minor typo in scheduler (#8765) 2024-09-25 00:53:03 -07:00
distributed [Core] Allow IPv6 in VLLM_HOST_IP with zmq (#8575) 2024-09-23 12:15:41 -07:00
engine [Model] Add support for the multi-modal Llama 3.2 model (#8811) 2024-09-25 13:29:32 -07:00
entrypoints [Model] Add support for the multi-modal Llama 3.2 model (#8811) 2024-09-25 13:29:32 -07:00
executor [Bugfix] Ray 2.9.x doesn't expose available_resources_per_node (#8767) 2024-09-25 00:52:26 -07:00
inputs [Model] Add support for the multi-modal Llama 3.2 model (#8811) 2024-09-25 13:29:32 -07:00
logging [MISC] Rework logger to enable pythonic custom logging configuration to be provided (#4273) 2024-05-01 17:34:40 -07:00
lora [Kernel][LoRA] Add assertion for punica sgmv kernels (#7585) 2024-09-23 18:57:42 +00:00
model_executor [Misc] Support quantization of MllamaForCausalLM (#8822) 2024-09-25 14:46:22 -07:00
multimodal [Model] Add support for the multi-modal Llama 3.2 model (#8811) 2024-09-25 13:29:32 -07:00
platforms [CI/Build] Avoid CUDA initialization (#8534) 2024-09-18 10:38:11 +00:00
plugins [plugin][torch.compile] allow to add custom compile backend (#8445) 2024-09-13 09:32:42 -07:00
prompt_adapter [CI/Build] Avoid CUDA initialization (#8534) 2024-09-18 10:38:11 +00:00
spec_decode [Core][Bugfix] Support prompt_logprobs returned with speculative decoding (#8047) 2024-09-24 17:29:56 -07:00
transformers_utils [Model] Add support for the multi-modal Llama 3.2 model (#8811) 2024-09-25 13:29:32 -07:00
triton_utils [CI/Build] Update Ruff version (#8469) 2024-09-18 11:00:56 +00:00
usage [CI/Build] Avoid CUDA initialization (#8534) 2024-09-18 10:38:11 +00:00
vllm_flash_attn [ci][build] fix vllm-flash-attn (#8699) 2024-09-21 23:24:58 -07:00
worker [Model] Add support for the multi-modal Llama 3.2 model (#8811) 2024-09-25 13:29:32 -07:00
__init__.py Revert "rename PromptInputs and inputs with backward compatibility (#8760) (#8810) 2024-09-25 10:36:26 -07:00
_core_ext.py [Bugfix] Allow ScalarType to be compiled with pytorch 2.3 and add checks for registering FakeScalarType and dynamo support. (#7886) 2024-08-27 23:13:45 -04:00
_custom_ops.py [Kernel] Fullgraph and opcheck tests (#8479) 2024-09-25 08:35:52 -06:00
_ipex_ops.py [Hardware][intel GPU] bump up ipex version to 2.3 (#8365) 2024-09-13 16:54:34 -07:00
block.py [mypy] Enable mypy type checking for vllm/core (#7229) 2024-08-28 07:11:14 +08:00
config.py [Model] Add support for the multi-modal Llama 3.2 model (#8811) 2024-09-25 13:29:32 -07:00
connections.py [core][distributed] fix zmq hang (#6759) 2024-07-24 17:37:12 -07:00
envs.py [misc] soft drop beam search (#8763) 2024-09-24 15:48:39 -07:00
logger.py [Bugfix] Don't disable existing loggers (#7664) 2024-08-19 15:11:58 -07:00
outputs.py Add output streaming support to multi-step + async while ensuring RequestOutput obj reuse (#8335) 2024-09-23 15:38:04 -07:00
pooling_params.py [Core] Optimize SPMD architecture with delta + serialization optimization (#7109) 2024-08-18 17:57:20 -07:00
py.typed Add py.typed so consumers of vLLM can get type checking (#1509) 2023-10-30 14:50:47 -07:00
sampling_params.py [misc] soft drop beam search (#8763) 2024-09-24 15:48:39 -07:00
scalar_type.py [Misc] Disambiguate quantized types via a new ScalarType (#6396) 2024-08-02 13:51:58 -07:00
scripts.py [BugFix] Fix clean shutdown issues (#8492) 2024-09-16 09:33:46 -07:00
sequence.py [Model] Add support for the multi-modal Llama 3.2 model (#8811) 2024-09-25 13:29:32 -07:00
tracing.py [CI/Build] Pin OpenTelemetry versions and make errors clearer (#7266) 2024-08-20 10:02:21 -07:00
utils.py [Core] Allow IPv6 in VLLM_HOST_IP with zmq (#8575) 2024-09-23 12:15:41 -07:00
version.py [CI/Build] use setuptools-scm to set __version__ (#4738) 2024-09-23 09:44:26 -07:00