The installation process is identical to MoSca:
-
Simply run the following command. This script assumes you have an Ubuntu environment and Anaconda installed. The CUDA version used is 11.8. You may need to tweak the script to fit your environment.
bash install.sh
-
Download from here the checkpoints for 2D foundational models if they are not Hugging Face downloadable.
WARNING: By downloading these checkpoints, you must agree and obey the original license from the original authors (RAFT, SpaTracker, and TAPNet). Unzip the weights into the following file structure:
ProjRoot/weights ├── raft_models │ ├── raft-things.pth │ └── ... ├── spaT_final.pth └── tapnet └── bootstapir_checkpoint_v2.pt
To reconstruct your own video, organize your data in the following structure:
demo/your_scene_name/
└── images/
├── 00000.jpg # or .png
├── 00001.jpg
├── 00002.jpg
└── ...Requirements:
- Image format:
.jpgor.png - Naming convention: Sequential numbering with zero-padding (e.g.,
00000.jpg,00001.jpg, ...) - Frame rate: 20-30 FPS recommended for dynamic scenes
- Resolution: The code will automatically resize images if needed
- Video extraction: If you have a video file, extract frames using:
ffmpeg -i your_video.mp4 -qscale:v 2 demo/your_scene_name/images/%05d.jpg
We provide demo scenes in the demo/ directory for quick testing. Each scene follows the structure above with images under demo/scene_name/images/.
python origs_precompute.py \
--cfg ./profile/demo/demo_prep.yaml \
--ws ./demo/luciaThis step runs off-the-shelf 2D foundational models (depth estimation, optical flow, tracking) on the input images.
python origs_reconstruct.py \
--cfg ./profile/demo/demo_fit.yaml \
--ws ./demo/luciaThis step performs the OriGS reconstruction pipeline.
You can also run both steps using the provided script:
bash demo.shdemo_prep.yaml: Controls 2D prior computation (depth model, tracking model, optical flow settings)demo_fit.yaml: Controls reconstruction pipeline (bundle adjustment parameters, scaffold settings, photometric optimization)
This work builds upon MoSca. If you use this code, please cite both OriGS and MoSca:
@inproceedings{wu2025orientation,
title={Orientation-anchored Hyper-Gaussian for 4D Reconstruction from Casual Videos},
author={Wu, Junyi and Tao, Jiachen and Wang, Haoxuan and Liu, Gaowen and Kompella, Ramana Rao and Yan, Yan},
booktitle={NeurIPS},
year={2025}
}
@inproceedings{lei2025mosca,
title={Mosca: Dynamic gaussian fusion from casual videos via 4d motion scaffolds},
author={Lei, Jiahui and Weng, Yijia and Harley, Adam W and Guibas, Leonidas and Daniilidis, Kostas},
booktitle={CVPR},
year={2025}
}