Official Pytorch implemetation of ICML2022 paper Depo (Plan Your Target and Learn Your Skills: Transferable State-Only Imitation Learning via Decoupled Policy Optimization) on NGSIM driving dataset. Since some settings of the driving environment are quite different from the Mujoco environments, we maintain the experiment code on NGSIM separately in this repo. The code for the main experiments can be found in DePO. Current simulation platform is a (slightly) modified version of PPUU. We plan to open source the training code based on the SMARTS simulator, so be sure to stay tuned!
Important Notes
This repository is based on ILSwiss. If you are looking for DePO for Mujoco experiments, check here.
Go to this address and download the TGZ file (330 MB) on your machine. Place the downloaded folder as pytorch-PPUU/traffic-data
.
Then transform the dataset so that it can fit in ILSwiss.
cd ILSwiss/
python run_experiment.py -e exp_specs/expert_ppuu.yaml
python run_experiment.py -e exp_specs/expert_ppuu_multitype.yaml
python run_experiment.py -e exp_specs/dpo_ppuu_multitype.yaml
- Pretrain on Normal
python run_experiment.py -e exp_specs/dpo_ppuu_pretrain.yaml
- Transfer to Inverse and Transpose
Remember to change state_predictor_path
in exp_specs/dpo_ppuu_transfer.yaml
to the path of best.pkl
file obtained from pretraining.
python run_experiment.py -e exp_specs/dpo_ppuu_transfer.yaml
Run GAIfO.
python run_experiment.py -e exp_specs/gailfo_ppuu.yaml