This is the official codebase for the Sirius paper:
Robot Learning on the Job: Human-in-the-Loop Autonomy and Learning During Deployment
Huihan Liu, Soroush Nasiriany, Lance Zhang, Zhiyao Bao, Yuke Zhu
UT Austin Robot Perception and Learning Lab
Robotics: Science and Systems (RSS), 2023
[Paper] [Project Website] [Real Robot Control]
Sirius builds upon robomimic, a framework for robot learning from demonstration. Sirius also uses the robotics simulator robosuite powered by the MuJoCo physics engine.
git clone https://github.com/UT-Austin-RPL/sirius
cd sirius
conda env create -f sirius.yml
conda activate sirius
pip install -e .
The additional reference for installing robosuite here and here could be helpful.
$ git clone https://github.com/ARISE-Initiative/robosuite.git
$ cd robosuite
$ pip install -r requirements.txt
$ pip install -e .
Running Sirius intervention-guided policy learning:
python robomimic/scripts/train.py --config robomimic/exps/sirius/sirius.json
IWR baseline:
python robomimic/scripts/train.py --config robomimic/exps/sirius/bc_iwr.json
BC baseline:
python robomimic/scripts/train.py --config robomimic/exps/sirius/bc.json
We include the script for collecting demonstrations and performing human intervention during robot policy execution below. We use a spacemouse for providing both demonstration and intervention. More details for setting up Spacemouse can be found here.
Perform human demonstration with the flag --all-demos
:
python robomimic/scripts/hitl/collect_hitl_demos.py --all-demos --num-traj 50
Perform human intervention with the policy checkpoint ${checkpoint}
:
python robomimic/scripts/hitl/collect_hitl_demos.py --num-traj 50 --checkpoint ${checkpoint}
By default, the datasets are generated in the minimum format with only low-level state information to save space. To add image observation and other modalities for training, run the following post-processing script. It will process the original data ${data.hdf5}
into ${data_processed.hdf5}
, with image size ${image_size}
. By default, the two camera view uses are agentview and robot0_eye_in_hand, which you can modify in the script template_process_sim_dataset.sh
.
cd robomimic/scripts/hitl
source template_process_sim_dataset.sh ${data.hdf5} ${data_processed.hdf5} ${image_size}
This codebase is largely built on robomimic and robosuite. We also thank Ajay Mandlekar for sharing well-designed simulation task environments beyond the robomimic codebase like Coffee
and Threading
tasks during project development.
For real-robot experiments, we used Deoxys, a controller library for Franka Emika Panda developed by Yifeng Zhu.
@inproceedings{liu2022robot,
title = {Robot Learning on the Job: Human-in-the-Loop Autonomy and Learning During Deployment},
author = {Huihan Liu and Soroush Nasiriany and Lance Zhang and Zhiyao Bao and Yuke Zhu},
booktitle = {Robotics: Science and Systems (RSS)},
year = {2023}
}