vllm/vllm/v1
Cody Yu 201fc07730
[V1] Prefix caching (take 2) (#9972)
Signed-off-by: Cody Yu <hao.yu.cody@gmail.com>
2024-11-07 17:34:44 -08:00
..
attention [V1] Make v1 more testable (#9888) 2024-11-06 11:57:35 -08:00
core [V1] Prefix caching (take 2) (#9972) 2024-11-07 17:34:44 -08:00
engine [V1] Prefix caching (take 2) (#9972) 2024-11-07 17:34:44 -08:00
executor [V1] Fix Configs (#9971) 2024-11-04 00:24:40 +00:00
sample [V1] Support per-request seed (#9945) 2024-11-03 09:14:17 -08:00
tokenizer [V1] Make v1 more testable (#9888) 2024-11-06 11:57:35 -08:00
worker [V1][BugFix] Fix Generator construction in greedy + seed case (#10097) 2024-11-07 05:06:57 +00:00
outputs.py [V1] Implement vLLM V1 [1/N] (#9289) 2024-10-22 01:24:07 -07:00
request.py [V1] Add all_token_ids attribute to Request (#10135) 2024-11-07 17:08:24 -08:00
utils.py [V1] Add all_token_ids attribute to Request (#10135) 2024-11-07 17:08:24 -08:00