| .. |
|
__init__.py
|
Change the name to vLLM (#150)
|
2023-06-17 03:07:40 -07:00 |
|
cache_engine.py
|
[Model] Support Mamba (#6484)
|
2024-10-11 15:40:06 +00:00 |
|
cpu_enc_dec_model_runner.py
|
[Hardware][CPU] Cross-attention and Encoder-Decoder models support on CPU backend (#9089)
|
2024-10-07 06:50:35 +00:00 |
|
cpu_model_runner.py
|
[Misc] Standardize RoPE handling for Qwen2-VL (#9250)
|
2024-10-16 13:56:17 +08:00 |
|
cpu_worker.py
|
[Model] Support Mamba (#6484)
|
2024-10-11 15:40:06 +00:00 |
|
embedding_model_runner.py
|
[Model] PP support for embedding models and update docs (#9090)
|
2024-10-06 16:35:27 +08:00 |
|
enc_dec_model_runner.py
|
[Model] Support Mamba (#6484)
|
2024-10-11 15:40:06 +00:00 |
|
model_runner_base.py
|
[MISC] Skip dumping inputs when unpicklable (#8744)
|
2024-09-24 06:10:03 +00:00 |
|
model_runner.py
|
[Core] Deprecating block manager v1 and make block manager v2 default (#8704)
|
2024-10-17 11:38:15 -05:00 |
|
multi_step_model_runner.py
|
[Doc] Compatibility matrix for mutual exclusive features (#8512)
|
2024-10-11 11:18:50 -07:00 |
|
multi_step_tpu_worker.py
|
[TPU] Implement multi-step scheduling (#8489)
|
2024-09-14 16:58:31 -07:00 |
|
multi_step_worker.py
|
[Core] Multi-Step + Single Step Prefills via Chunked Prefill code path (#8378)
|
2024-09-27 13:32:07 -07:00 |
|
neuron_model_runner.py
|
[Core][Frontend] Add Support for Inference Time mm_processor_kwargs (#9131)
|
2024-10-08 14:12:56 +00:00 |
|
neuron_worker.py
|
[Bugfix] neuron: enable tensor parallelism (#7562)
|
2024-08-26 15:13:13 -07:00 |
|
openvino_model_runner.py
|
[Model] Support Mamba (#6484)
|
2024-10-11 15:40:06 +00:00 |
|
openvino_worker.py
|
[Model] Support Mamba (#6484)
|
2024-10-11 15:40:06 +00:00 |
|
tpu_model_runner.py
|
[TPU] Fix TPU SMEM OOM by Pallas paged attention kernel (#9350)
|
2024-10-14 15:02:06 -07:00 |
|
tpu_worker.py
|
[torch.compile] use empty tensor instead of None for profiling (#8875)
|
2024-09-27 08:11:32 -07:00 |
|
utils.py
|
[Doc] Compatibility matrix for mutual exclusive features (#8512)
|
2024-10-11 11:18:50 -07:00 |
|
worker_base.py
|
[Core] Logprobs support in Multi-step (#7652)
|
2024-08-29 19:19:08 -07:00 |
|
worker.py
|
[Bugfix][Core] Use torch.cuda.memory_stats() to profile peak memory usage (#9352)
|
2024-10-17 22:47:27 -04:00 |
|
xpu_model_runner.py
|
[Hardware][intel GPU] add async output process for xpu (#8897)
|
2024-10-14 12:23:33 -06:00 |
|
xpu_worker.py
|
[Hardware][Intel GPU] Add intel GPU pipeline parallel support. (#7810)
|
2024-08-27 10:07:02 -07:00 |