Skip to content
This repository was archived by the owner on Nov 1, 2024. It is now read-only.

facebookresearch/ssorl

ssorl

This repository contains the Pytorch implementation of Semi-Supervised Offline Reinforcement Learning with Action-Free Trajectories by Qinqing Zheng, Mikael Henaff, Brandon Amos, and Aditya Grover.

If you use this code for your research, please cite us as:

@inproceedings{zheng2023semi,
  title={Semi-supervised offline reinforcement learning with action-free trajectories},
  author={Zheng, Qinqing and Henaff, Mikael and Amos, Brandon and Grover, Aditya},
  booktitle={International Conference on Machine Learning},
  pages={42339--42362},
  year={2023},
  organization={PMLR}
}

Requirements

Install the conda environment:

conda env create -f conda_env.yml
conda activate ssorl

Update the path:

export LD_LIBRARY_PATH=$LD_LIBRARY_PATH:<your_conda_env_path>/lib

If you haven't installed patchelf, run:

sudo apt-get install patchelf

Example

Run the following command to train a SS-TD3BC agent for hopper with the medium-v2 dataset, where 10% trajecotories, whose returns are from the lower 50%, contain actions.

python main.py

This will produce the exp-local folder, where all the outputs are going to be logged including tensorboard blobs. One can attach a tensorboard to monitor training by running:

tensorboard --logdir exp-local

License

The majority of ssorl is licensed under CC-BY-NC, however portions of the project are available under separate license terms:

  • D4RL dataset - Creative Commons Attribution 4.0 License (CC-BY)
  • D4RL code, transformers, Lamb - Apache 2.0 License
  • stable-baselines3, Gym, decision-transformer - MIT License

About

Semi-Supervised Offline Reinforcement Learning with Action-Free Trajectories

Resources

License

Code of conduct

Contributing

Security policy

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages