vllm/tests
Tyler Michael Smith 85657b5607
[Kernel] Factor out epilogues from cutlass kernels (#5391)
Co-authored-by: Michael Goin <michael@neuralmagic.com>
Co-authored-by: youkaichao <youkaichao@gmail.com>
Co-authored-by: zifeitong <zifei.tong@parasail.io>
Co-authored-by: Robert Shaw <114415538+robertgshaw2-neuralmagic@users.noreply.github.com>
2024-06-13 11:22:19 -07:00
..
async_engine [CI/Build] Simplify OpenAI server setup in tests (#5100) 2024-06-13 11:21:53 -07:00
basic_correctness [CI/Test] improve robustness of test (vllm_runner) (#5357) 2024-06-08 08:59:20 +00:00
core [CI] Upgrade codespell version. (#5381) 2024-06-12 10:06:14 -07:00
distributed [Core][Distributed] code deduplication in tp&pp with coordinator(#5293) 2024-06-12 17:27:08 -07:00
engine [CI/Test] improve robustness of test (vllm_runner) (#5357) 2024-06-08 08:59:20 +00:00
entrypoints [CI/Build] Simplify OpenAI server setup in tests (#5100) 2024-06-13 11:21:53 -07:00
fp8_kv Enable scaled FP8 (e4m3fn) KV cache on ROCm (AMD GPU) (#3290) 2024-04-03 14:15:55 -07:00
kernels [Kernel] Factor out epilogues from cutlass kernels (#5391) 2024-06-13 11:22:19 -07:00
lora [Core][Distributed] code deduplication in tp&pp with coordinator(#5293) 2024-06-12 17:27:08 -07:00
metrics [CI/Test] improve robustness of test (vllm_runner) (#5357) 2024-06-08 08:59:20 +00:00
model_executor [CI/Build] Move test_utils.py to tests/utils.py (#4425) 2024-05-13 23:50:09 +09:00
models [CI/Build][REDO] Add is_quant_method_supported to control quantization test configurations (#5466) 2024-06-13 15:18:08 +00:00
multimodal [Model] Initial support for LLaVA-NeXT (#4199) 2024-06-10 12:47:15 +00:00
prefix_caching [Bugfix / Core] Prefix Caching Guards (merged with main) (#4846) 2024-05-27 15:18:17 -07:00
prompts [BugFix] Fix input positions for long context with sliding window (#2088) 2023-12-13 12:28:13 -08:00
quantization [CI/Build][REDO] Add is_quant_method_supported to control quantization test configurations (#5466) 2024-06-13 15:18:08 +00:00
samplers [CI/Test] improve robustness of test (vllm_runner) (#5357) 2024-06-08 08:59:20 +00:00
spec_decode [Core][Doc] Default to multiprocessing for single-node distributed case (#5230) 2024-06-11 11:10:41 -07:00
tensorizer_loader [CI/Build] Simplify OpenAI server setup in tests (#5100) 2024-06-13 11:21:53 -07:00
tokenization [Core] Support image processor (#4197) 2024-06-02 22:56:41 -07:00
worker [Core][Distributed] code deduplication in tp&pp with coordinator(#5293) 2024-06-12 17:27:08 -07:00
__init__.py [Small] Formatter only checks lints in changed files (#1528) 2023-10-31 15:39:38 -07:00
conftest.py [Core][Distributed] code deduplication in tp&pp with coordinator(#5293) 2024-06-12 17:27:08 -07:00
test_cache_block_hashing.py [Core] Avoid the need to pass None values to Sequence.inputs (#5099) 2024-05-29 16:05:01 -07:00
test_config.py [Frontend] Customizable RoPE theta (#5197) 2024-06-11 10:42:26 -07:00
test_inputs.py [Core] Consolidate prompt arguments to LLM engines (#4328) 2024-05-28 13:29:31 -07:00
test_logger.py [MISC] Rework logger to enable pythonic custom logging configuration to be provided (#4273) 2024-05-01 17:34:40 -07:00
test_logits_processor.py [Misc] Remove unnecessary ModelRunner imports (#4703) 2024-05-09 00:17:17 -07:00
test_regression.py Bugfix: fix broken of download models from modelscope (#5233) 2024-06-06 09:28:10 -07:00
test_sampling_params.py [Bugfix] fix crash if max_tokens=None (#2570) 2024-01-23 22:38:55 -08:00
test_sequence.py [CI/Build] Move test_utils.py to tests/utils.py (#4425) 2024-05-13 23:50:09 +09:00
test_sharded_state_loader.py [CI] Upgrade codespell version. (#5381) 2024-06-12 10:06:14 -07:00
test_utils.py [Misc][Utils] allow get_open_port to be called for multiple times (#5333) 2024-06-06 22:15:11 -07:00
utils.py [CI/Build] Simplify OpenAI server setup in tests (#5100) 2024-06-13 11:21:53 -07:00