From ee8217e5bee5860469204ee57077a91138c9af02 Mon Sep 17 00:00:00 2001 From: amaleshvemula Date: Wed, 11 Oct 2023 09:26:24 +0200 Subject: [PATCH] Add Mistral to quantization model list (#1278) --- vllm/model_executor/model_loader.py | 1 + 1 file changed, 1 insertion(+) diff --git a/vllm/model_executor/model_loader.py b/vllm/model_executor/model_loader.py index 951ba1f0..82c8dfa7 100644 --- a/vllm/model_executor/model_loader.py +++ b/vllm/model_executor/model_loader.py @@ -35,6 +35,7 @@ _MODEL_REGISTRY = { # FIXME(woosuk): Remove this once all models support quantization. _MODEL_CLASSES_SUPPORT_QUANTIZATION = [ LlamaForCausalLM, + MistralForCausalLM, ]