Code accompanying the paper "Action Robust Reinforcement Learning and Applications in Continuous Control"
- MuJoCo
- Python 3 (it might work with Python 2, not tested)
- PyTorch
- OpenAI Gym
- tdqm
- numpy
- matplotlib
python3.6 main.py --updates_per_step 10 --env-name "Hopper-v2" --alpha 0.1 --method pr_mdp
Where method can take 3 values mdp
pr_mdp
or nr_mdp
, where pr/nr
are the probabilistic robust and noisy robust as defined in the paper.
All results are saved in the models folder.
Once a model has been trained, run:
python3.6 test.py --eval_type model
where --eval_type model
will evaluate for model (mass) uncertainty and --eval_type model_noise
will create the 2d visualizations.
See Comparison_Plots.ipynb
for an example of how to access and visualize your models.