From 8a1415cf776b2b902f6429ecfc325877b57cbefe Mon Sep 17 00:00:00 2001 From: Thomas Parnell Date: Thu, 11 Jul 2024 16:05:59 +0200 Subject: [PATCH] [Bugfix] GPTBigCodeForCausalLM: Remove lm_head from supported_lora_modules. (#6326) Signed-off-by: Thomas Parnell Co-authored-by: Travis Johnson --- vllm/model_executor/models/gpt_bigcode.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/vllm/model_executor/models/gpt_bigcode.py b/vllm/model_executor/models/gpt_bigcode.py index cc42413d..fc4e13bb 100644 --- a/vllm/model_executor/models/gpt_bigcode.py +++ b/vllm/model_executor/models/gpt_bigcode.py @@ -235,7 +235,7 @@ class GPTBigCodeModel(nn.Module): class GPTBigCodeForCausalLM(nn.Module, SupportsLoRA): packed_modules_mapping = {"c_attn": ["c_attn"]} - supported_lora_modules = ["c_fc", "c_proj", "wte", "lm_head", "c_attn"] + supported_lora_modules = ["c_fc", "c_proj", "wte", "c_attn"] embedding_modules = { "wte": "input_embeddings",