Skip to content

jiaming-zhou/HumanRobotAlign

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

5 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

[CVPR 2025] Mitigating the Human-Robot Domain Discrepancy in Visual Pre-training for Robotic Manipulation

Jiaming Zhou1, Teli Ma1, Kun-Yu Lin2, Zifan Wang1, Ronghe Qiu1, Junwei Liang1,3

1AI Thrust, HKUST (Guangzhou), 2Sun Yat-sen University, 3HKUST

Overview

This is the official repo for HumanRobotAlign, which proposes a novel adaptation paradigm that leverages readily available paired human-robot video data to bridge the domain gap between human-data pre-trained models and downstream robotic manipulation tasks.

img

Our adaptation paradigm includes two stages.

  • The first adaptation stage adapts existing human-data pre-trained models using semantics-aligned human-robot video pairs.
  • The second evaluation stage evaluates the adapted models on downstream manipulation benchmarks. The evaluation includes downstream-training part and downstream-test part. The downstream-training part utilizes the adapted pre-trained models as fronzen backbones for policy learning. And the downstream-test part tests the learned policies.

If you find this useful, please cite the paper!

@article{zhou2024mitigating,
  title={Mitigating the human-robot domain discrepancy in visual pre-training for robotic manipulation},
  author={Zhou, Jiaming and Ma, Teli and Lin, Kun-Yu and Wang, Zifan and Qiu, Ronghe and Liang, Junwei},
  journal={arXiv preprint arXiv:2406.14235},
  year={2024}
}

Evaluation Stage on RLBench Benchmark.

Installation

1. Clone this repo

git clone [email protected]:jiaming-zhou/HumanRobotAlign.git

2. Install conda environment and simulator

Following the RVT to install conda environment and CoppeliaSim.

3. Install RVT, PyRep, RLBench, YARR and PerAct Colab

cd HumanRobotAlign
pip install -e .
pip install -e rvt/libs/PyRep 
pip install -e rvt/libs/RLBench 
pip install -e rvt/libs/YARR 
pip install -e rvt/libs/peract_colab

Training Downstream Policies

Download the RLBench Replay data for training provided by RVT.

Set the value of TRAIN_REPLAY_STORAGE_DIR in "rvt/train.py" to the rlbench replay data path.

[UnadaptedR3M2RLBench]: using R3M's pre-trained model as frozen backbone to train policy on RLBench:

  • download R3M's pre-trained model (UnadaptedR3M.pt) and put it under subfolder 'pretrains';
  • then run the following command:
    cd rvt
    python train.py --exp_cfg_path configs/unadaptedR3M.yaml --device [gpu_ids]
    

[AdaptedR3M2RLBench]: using our adapted R3M model as frozen backbone to train policy on RLBench:

  • download our adapted R3M model (AdaptedR3M.pyth) and put it under subfolder 'pretrains';
  • then run the following command:
    cd rvt
    python train.py --exp_cfg_path configs/adaptedR3M.yaml --device [gpu_ids]
    

Test Downstream Policies

Download the RLBench Test data for testing by following PerAct repo.

Test [UnadaptedR3M2RLBench]:

  • download our trained UnadaptedR3M2RLBench policy and unzip it under the subfolder 'rvt/runs', or use the policy trained by yourself;
  • then run the following command:
    cd rvt
    xvfb-run -a python eval.py --model-folder runs/UnadaptedR3M2RLBench --eval-datafolder /your_data_path/RLBench/test --tasks all --eval-episodes 25 --log-name your_logname --device 0 --headless --model-name model_4.pth
    

Test [AdaptedR3M2RLBench]:

  • download our trained AdaptedR3M2RLBench policy and unzip it under the subfolder 'rvt/runs', or use the policy trained by yourself;
  • then run the following command:
    cd rvt
    xvfb-run -a python eval.py --model-folder runs/AdaptedR3M2RLBench --eval-datafolder /your_data_path/RLBench/test --tasks all --eval-episodes 25 --log-name your_logname --device 0 --headless --model-name model_4.pth
    

Acknowledgments

RLBench, PerAct, RVT.

Bibtex

If you find this useful, please cite the paper!

@article{zhou2024mitigating,
  title={Mitigating the human-robot domain discrepancy in visual pre-training for robotic manipulation},
  author={Zhou, Jiaming and Ma, Teli and Lin, Kun-Yu and Wang, Zifan and Qiu, Ronghe and Liang, Junwei},
  journal={arXiv preprint arXiv:2406.14235},
  year={2024}
}

About

This is the official repo for [CVPR 2025] paper, Mitigating the Human-Robot Domain Discrepancy in Visual Pre-training for Robotic Manipulation. https://jiaming-zhou.github.io/projects/HumanRobotAlign/

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages