From 091236ad34667b27f9915907e2d0295c42893a0b Mon Sep 17 00:00:00 2001 From: qianyu chen <38046403+qyc-98@users.noreply.github.com> Date: Wed, 17 Jul 2024 14:37:18 +0800 Subject: [PATCH] Update finetune_lora.sh --- finetune/finetune_lora.sh | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/finetune/finetune_lora.sh b/finetune/finetune_lora.sh index 49c375b..22cf5a2 100644 --- a/finetune/finetune_lora.sh +++ b/finetune/finetune_lora.sh @@ -37,7 +37,7 @@ torchrun $DISTRIBUTED_ARGS finetune.py \ --tune_vision true \ --tune_llm false \ --use_lora true \ - --lora_target_modules "llm\..*layers\.\d+\.self_attn\.(q_proj|k_proj)" \ + --lora_target_modules "llm\..*layers\.\d+\.self_attn\.(q_proj|k_proj|v_proj|o_proj)" \ --model_max_length 2048 \ --max_slice_nums 9 \ --max_steps 10000 \