CUDA_VISIBLE_DEVICES=1 python3.12 train_stage2_v0.py \
--model_name_or_path="/datas/MIRAS-7B_2" \
--dataset_dir='/datas/multimodal_datasets/GRESDataset' \
--log_base_dir='/datas/runs' \
--vision_pretrained="/datas/sam_vit_h_4b8939.pth" \
--dataset="multiturn||vqa" \
--vqa_data="filtered_sharegpt4v" \
--val_dataset="Multiturn|val" \
--sample_rates="10,2" \
--exp_name="mgmsa_stage2_5_noobj" \
--refer_seg_data="refclef||refcoco||refcoco+||refcocog" \
--batch_size=8 \
--pretrain_mm_mlp_adapter="/datas/mm_7b_projector.bin" \
--image_size_aux=768 \
--lora_r=8 \
--epochs=100 \
--steps_per_epoch=250 \cd /datas/MGM_/runs/miras/ckpt_model && python zero_to_fp32.py . ../pytorch_model.binCUDA_VISIBLE_DEVICES=0 python /datas/MGM_/merge_lora_weights_and_save_hf_model.py \
--version="/datas/llava-v1.5-7b" \
--weight="/datas/MGM_/runs/miras/pytorch_model.bin" \
--save_path="MIRAS" \
--lora_r=8 \deepspeed --master_port=20999 --include localhost:1
CUDA_VISIBLE_DEVICES=1 python3.12 train_ds_7b.py \
--model_name_or_path='/datas/MIRAS' \
--vision_pretrained='/datas/sam_vit_h_4b8939.pth' \
--dataset_dir='/datas/multimodal_datasets/Dataset' \ //数据集总路径
--dataset='reason_seg||caption' \
--lora_r=8 \
--pretrain_mm_mlp_adapter="/datas/mm_7b_projector.bin" \
--image_size_aux=768 \
--val_dataset="refcocog|umd|val" \
--eval_only \CUDA_VISIBLE_DEVICES=3 python chat.py --version='/datas/MIRAS-7B'