diff --git a/finetune/finetune.py b/finetune/finetune.py index 671d6c7..d2d2cff 100644 --- a/finetune/finetune.py +++ b/finetune/finetune.py @@ -260,6 +260,7 @@ def train(): lora_dropout=lora_args.lora_dropout, bias=lora_args.lora_bias, layers_to_transform=lora_args.lora_layers_to_transform, + modules_to_save=modules_to_save, ) if not hasattr(model, 'get_input_embeddings'): def get_input_embeddings(self):