MinD-3D: Reconstruct High-quality 3D objects in Human Brain
Jianxiong Gao, Yuqian Fu, Yun Wang, Xuelin Qian, Jianfeng Feng, Yanwei Fu†
ECCV, 2024
MinD-3D++: Advancing fMRI-Based 3D Reconstruction with High-Quality Textured Mesh Generation and a Comprehensive Dataset
Jianxiong Gao, Yanwei Fu†, Yuqian Fu, Yun Wang, Xuelin Qian, Jianfeng Feng
TPAMI, 2025
- [08/2025] MinD-3D++ is accepted by TPAMI!
- [08/2025] We have released the code for MinD-3D++!
- [11/2024] We have released the training code for MinD-3D!
You can download fMRI-Shape by this link: https://huggingface.co/datasets/Fudan-fMRI/fMRI-Shape.
You can download fMRI-Objaverse by this link: https://huggingface.co/datasets/Fudan-fMRI/fMRI-Objaverse.
git clone https://github.com/JianxGao/MinD-3D.git
cd MinD-3D
bash env_install.shCUDA_VISIBLE_DEVICES=0 python -m torch.distributed.launch --nproc_per_node=1 --master_port=25645 \
train_stage1.py --sub_id 0001 --ddp \
--config ./configs/mind3d.yaml \
--out_dir sub01_stage1 --batchsize 8CUDA_VISIBLE_DEVICES=1 python -m torch.distributed.launch --nproc_per_node=1 --master_port=25645 \
train_stage2.py --sub_id 0001 --ddp \
--config ./configs/mind3d.yaml \
--out_dir sub01_stage2 --batchsize 2You can access the quantized features for training through the link: https://drive.google.com/file/d/1R8IpG1bligLAfHkLQ2COrfTIkay14AEm/view?usp=drive_link.
You can download the weight of subject 1 through the link: https://drive.google.com/file/d/1ni4g1iCvdpoi2xYtmydr_w3XA5PpNrvm/view?usp=sharing
# Sub01 Plane
python generate_fmri2shape.py --config ./configs/mind3d.yaml --check_point_path ./mind3d_sub01.pt \
--uid b5d0ae4f723bce81f119374ee5d5f944 --topk 250
# Sub01 Car
python generate_fmri2shape.py --config ./configs/mind3d.yaml --check_point_path ./mind3d_sub01.pt \
--uid aebd98c5d7e8150b709ce7955adef61b --topk 250For detailed instructions, please refer to the InstantMesh.
CUDA_VISIBLE_DEVICES=0,1,2,3 python -m torch.distributed.launch --nproc_per_node=4 --master_port=25644 \
python train_mind3d_pp.py --ddp \
--config ./configs/mind3d_pp.yaml \
--out_dir mind3dpp_fmri_shape_subject1_rank_64 --batchsize 8cd InstantMesh
CUDA_VISIBLE_DEVICES=0 python infer_fmri_obj.py ./configs/mind3d_pp_infer.yaml \
--unet_path model_weight \
--save_name save_dir \
--input_path ./dataset/fmri_shape/core_test_list.txt \
--fmri_dir fmri_dir \
--gt_image_dir gt_image_dir \
--save_video --export_texmapIf you would like to preprocess our data from scratch, please refer to the data_process folder and follow the steps below.
First, run the following command to preprocess the raw fMRI data using fMRIPrep. This step will generate dtseries data as output. (It may cost one day)
~/.local/bin/fmriprep-docker path/to/fMRI_Shape/dcm2bids/BIDS path/to/fMRI_Shape/dcm2bids/BIDS2 participant --skip_bids_validation \
--participant-label 0003 -w path/to/fMRI_Shape/dcm2bids/fmriprep_BIDS2/tmp --nthreads 32 --omp-nthreads 32 \
--output-spaces MNI152NLin6Asym:res-2 fsaverage5 --cifti-output --use-aroma --ignore slicetiming sbref t2w \
--fs-subjects-dir path/to/fMRI_Shape/dcm2bids/Test/freesurfer --fs-license-file path/to/gaojianxiong/license.txtAfter this step, the preprocessed CIFTI dtseries files will be generated.
If you want to obtain surface-based fMRI images in the same format as shown in our paper, you should first install Connectome Workbench:
After setting up Connectome Workbench, please run the following script:
python nii2surf_dnv.pyThis script converts the preprocessed dtseries data into surface-based fMRI visualizations that are consistent with those used in the paper.
After generating the surface-based fMRI data, run the following script to convert .h5 files into .npy format:
python h5tonpy.pyPlease make sure to modify the file paths in the script according to your local directory structure.
If you find our paper useful for your research and applications, please cite using this BibTeX:
@misc{gao2023mind3d,
title={MinD-3D: Reconstruct High-quality 3D objects in Human Brain},
author={Jianxiong Gao and Yuqian Fu and Yun Wang and Xuelin Qian and Jianfeng Feng and Yanwei Fu},
year={2023},
eprint={2312.07485},
archivePrefix={arXiv},
primaryClass={cs.CV}
}
@misc{gao2025mind3dadvancingfmribased3d,
title={MinD-3D++: Advancing fMRI-Based 3D Reconstruction with High-Quality Textured Mesh Generation and a Comprehensive Dataset},
author={Jianxiong Gao and Yanwei Fu and Yuqian Fu and Yun Wang and Xuelin Qian and Jianfeng Feng},
year={2025},
eprint={2409.11315},
archivePrefix={arXiv},
primaryClass={cs.CV},
url={https://arxiv.org/abs/2409.11315},
}



