Discriminator-Actor-Critic: Addressing Sample Inefficiency and Reward Bias in Adversarial Imitation Learning
Source code to accompany our paper.
We use Python 3.5.4rc1. You may also need to install a number of dependencies.
pip3 install gym
pip3 install --upgrade tensorflow tensorflow_probability
pip3 install absl-py
You will also need to install Mujoco and use a valid license. Follow the install instructions here.
Clone the repo of expert trajectories:
cd /data/dac/ # We will assume access to this directory.
git clone [email protected]:ikostrikov/gail-experts.git
Then use our import script to turn them into checkpoints (~1-2 hours):
python3 generate_expert_data.py \
--src_data_dir /data/dac/gail-experts/ \
--dst_data_dir /data/dac/gail-experts/
Launch run_training_worker.sh to start the training worker. Then in another terminal, launch run_evaluation_worker.sh. Training takes approximately 1 to 2 hours.
To change the environment, number of expert trajectories, etc, edit the variables defined in the bash scripts above.
To see reward results live during training, launch a tensorboard:
tensorboard --logdir /tmp/lfd_state_action_traj_4_HalfCheetah-v2_20