mirror of
https://github.com/OpenBMB/MiniCPM-V.git
synced 2026-02-05 10:19:18 +08:00
update lora finetune inference bug (#224)
This commit is contained in:
@@ -74,7 +74,7 @@ Specially, Llama3 has a different chat_template for training and inference, we m
|
||||
The LoRA allows light-weight model tuning with only a small subset of parameters updated. We provide the LoRA implementation based on `peft`. To launch your training, run the following script:
|
||||
|
||||
```
|
||||
sh finetune_ds_lora.sh
|
||||
sh finetune_lora.sh
|
||||
```
|
||||
|
||||
After training, you could load the model with the path to the adapter. We advise you to use absolute path for your pretrained model. This is because LoRA only saves the adapter and the absolute path in the adapter configuration json file is used for finding out the pretrained model to load.
|
||||
@@ -82,12 +82,18 @@ After training, you could load the model with the path to the adapter. We advise
|
||||
```
|
||||
from peft import AutoPeftModelForCausalLM
|
||||
|
||||
path_to_adapter="path_to_adapter"
|
||||
|
||||
model = AutoPeftModelForCausalLM.from_pretrained(
|
||||
# path to the output directory
|
||||
path_to_adapter,
|
||||
device_map="auto",
|
||||
trust_remote_code=True
|
||||
).eval()
|
||||
|
||||
vpm_resampler_embedtokens_weight = torch.load(f"{path_to_adapter}/vpm_resampler_embedtokens.pt")
|
||||
|
||||
msg = model.load_state_dict(vpm_resampler_embedtokens_weight, strict=False)
|
||||
```
|
||||
|
||||
|
||||
@@ -122,10 +128,6 @@ A:When you face Out of Memory (OOM) issues during training large models, the f
|
||||
```
|
||||
--batch_size 1
|
||||
```
|
||||
- **Lower image resolution**: If your model processes image data, reducing the input resolution of images can effectively decrease memory usage.
|
||||
```
|
||||
--scale_resolution 448
|
||||
```
|
||||
- **Reduce the number of slices (`slice`)**: When handling large datasets such as large images files, reducing the number of slices processed each time can lower memory requirements.
|
||||
```
|
||||
--max_slice_nums 9
|
||||
|
||||
Reference in New Issue
Block a user