mirror of
https://github.com/OpenBMB/MiniCPM-V.git
synced 2026-02-04 17:59:18 +08:00
Modify eval_mm for MiniCPM-V 2.6
This commit is contained in:
31
eval_mm/vlmevalkit/scripts/run_inference.sh
Normal file
31
eval_mm/vlmevalkit/scripts/run_inference.sh
Normal file
@@ -0,0 +1,31 @@
|
||||
export PATH=/usr/local/cuda/bin:$PATH
|
||||
|
||||
export HF_ENDPOINT=https://hf-mirror.com
|
||||
export OMP_NUM_THREADS=1
|
||||
export timestamp=`date +"%Y%m%d%H%M%S"`
|
||||
export OLD_VERSION='False'
|
||||
export PYTHONPATH=$(dirname $SELF_DIR):$PYTHONPATH
|
||||
|
||||
# gpu consumed
|
||||
# fp16 17-18G
|
||||
# int4 7-8G
|
||||
|
||||
# model to be used
|
||||
# Example: MODELNAME=MiniCPM_V_2_6
|
||||
MODELNAME=$1
|
||||
# datasets to be tested
|
||||
# Example: DATALIST="MMMU_DEV_VAL MathVista_MINI MMVet MMBench_DEV_EN_V11 MMBench_DEV_CN_V11 MMStar HallusionBench AI2D_TEST"
|
||||
DATALIST=$2
|
||||
# test mode, all or infer
|
||||
MODE=$3
|
||||
|
||||
echo "Starting inference with model $MODELNAME on datasets $DATALIST"
|
||||
# run on multi gpus with torchrun command
|
||||
# remember to run twice, the first run may fail
|
||||
torchrun --nproc_per_node=8 run.py --data $DATALIST --model $MODELNAME --mode $MODE
|
||||
torchrun --nproc_per_node=8 run.py --data $DATALIST --model $MODELNAME --mode $MODE
|
||||
# run on single gpu with python command
|
||||
# python run.py --data $DATALIST --model $MODELNAME --verbose --mode $MODE
|
||||
# python run.py --data $DATALIST --model $MODELNAME --verbose --mode $MODE
|
||||
|
||||
ls
|
||||
Reference in New Issue
Block a user