mirror of
https://github.com/OpenBMB/MiniCPM-V.git
synced 2026-02-05 18:29:18 +08:00
Update zero3 code and OOM FQAs (#188)
This commit is contained in:
@@ -50,6 +50,8 @@ class TrainingArguments(transformers.TrainingArguments):
|
||||
tune_llm: Optional[bool] = field(default=True)
|
||||
llm_type: str = field(default="minicpm")
|
||||
use_lora: Optional[bool] = field(default=False)
|
||||
max_slice_nums: Optional[int] = field(default=9)
|
||||
scale_resolution: Optional[int] = field(default=448)
|
||||
|
||||
|
||||
@dataclass
|
||||
@@ -272,11 +274,17 @@ def train():
|
||||
|
||||
rank0_print(f'llm_type={llm_type}')
|
||||
|
||||
|
||||
# Load data
|
||||
if hasattr(model.config, "slice_config"):
|
||||
model.config.slice_config.max_slice_nums = training_args.max_slice_nums
|
||||
model.config.slice_config.scale_resolution = training_args.scale_resolution
|
||||
slice_config = model.config.slice_config.to_dict()
|
||||
else:
|
||||
model.config.max_slice_nums = training_args.max_slice_nums
|
||||
model.config.scale_resolution = training_args.scale_resolution
|
||||
slice_config = model.config.to_dict()
|
||||
|
||||
if hasattr(model.config, "batch_vision_input"):
|
||||
batch_vision = model.config.batch_vision_input
|
||||
else:
|
||||
|
||||
Reference in New Issue
Block a user