From e1aa52f714d9ec925ea48848d2e4611e61ec64dc Mon Sep 17 00:00:00 2001 From: YuzaChongyi <490083538@qq.com> Date: Tue, 28 May 2024 17:11:48 +0800 Subject: [PATCH 1/2] Update finetune_ds.sh --- finetune/finetune_ds.sh | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/finetune/finetune_ds.sh b/finetune/finetune_ds.sh index 8d90f11..ecc6cd6 100644 --- a/finetune/finetune_ds.sh +++ b/finetune/finetune_ds.sh @@ -34,8 +34,8 @@ torchrun $DISTRIBUTED_ARGS finetune.py \ --fp16_full_eval false \ --do_train \ --do_eval \ - --tune_vision false \ - --tune_llm false \ + --tune_vision true \ + --tune_llm true \ --model_max_length 2048 \ --max_steps 10000 \ --eval_steps 1000 \ From ae0f1e63ee71bf5a43b87f67873319e56e5c1aa0 Mon Sep 17 00:00:00 2001 From: YuzaChongyi <490083538@qq.com> Date: Tue, 28 May 2024 17:12:24 +0800 Subject: [PATCH 2/2] Update finetune_lora.sh --- finetune/finetune_lora.sh | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/finetune/finetune_lora.sh b/finetune/finetune_lora.sh index be3fb38..deba0d5 100644 --- a/finetune/finetune_lora.sh +++ b/finetune/finetune_lora.sh @@ -32,7 +32,7 @@ torchrun $DISTRIBUTED_ARGS finetune.py \ --bf16_full_eval true \ --do_train \ --do_eval \ - --tune_vision false \ + --tune_vision true \ --tune_llm false \ --use_lora true \ --lora_target_modules "llm\..*layers\.\d+\.self_attn\.(q_proj|k_proj)" \