vllm/vllm/multimodal
Chen Zhang 770ec6024f
[Model] Add support for the multi-modal Llama 3.2 model (#8811)
Co-authored-by: simon-mo <xmo@berkeley.edu>
Co-authored-by: Chang Su <chang.s.su@oracle.com>
Co-authored-by: Simon Mo <simon.mo@hey.com>
Co-authored-by: Roger Wang <136131678+ywang96@users.noreply.github.com>
Co-authored-by: Roger Wang <ywang@roblox.com>
2024-09-25 13:29:32 -07:00
..
__init__.py [Core][VLM] Stack multimodal tensors to represent multiple images within each prompt (#7902) 2024-08-28 01:53:56 +00:00
audio.py [Frontend][Core] Add plumbing to support audio language models (#7446) 2024-08-13 17:39:33 +00:00
base.py [Model] Add support for the multi-modal Llama 3.2 model (#8811) 2024-09-25 13:29:32 -07:00
image.py [Model] Add support for the multi-modal Llama 3.2 model (#8811) 2024-09-25 13:29:32 -07:00
registry.py [Core][Frontend] Support Passing Multimodal Processor Kwargs (#8657) 2024-09-23 07:44:48 +00:00
utils.py [model] Support for Llava-Next-Video model (#7559) 2024-09-10 22:21:36 -07:00
video.py [Core][Frontend] Support Passing Multimodal Processor Kwargs (#8657) 2024-09-23 07:44:48 +00:00