Skip to content

This project contains various reinforcement/imitation learning experiments.

License

Notifications You must be signed in to change notification settings

liquidnode/RLexperiments

Repository files navigation

RLexperiments

This project contains various reinforcement/imitation learning experiments, such as implementations of Implicit Quantile Networks [1], Neural Episodic Control [2], FeUdal Networks [3], Bootstrapped Dual Policy Iteration [4], Prioritized Experience Replay [5], TD3 [6], MuZero [7]. Most of them are applied to the MineRL enviroment package. The two latest experiments, an off-policy multiaction actor-critic agent and an Upside Down RL agent [8, 9], are featured in the root directory of this repository. One resulting multiaction agent can be seen here https://www.youtube.com/watch?v=oyNKCeMywtY. The other experiments can be found in the all_experiments/ directory.

This project also features an easily expandable multi-processing system of producer/consumer workers (trajectory generator, replay buffer and trainer workers).

This repository contains modified and unmodified code from:

References:

[1] W. Dabney et al. Implicit quantile networks for distributional reinforcement learning. In International Conference on Machine Learning, pages 1104–1113, 2018.

[2] A. Pritzel et al. Neural episodic control. In Proceedings of the 34th International Conference on Machine Learning-Volume 70, pages 2827–2836. JMLR. org, 2017.

[3] A. S. Vezhnevets et al. Feudal networks for hierarchical reinforcement learning. In Proceedings of the 34th International Conference on Machine Learning-Volume 70, pages 3540–3549. JMLR. org, 2017.

[4] D. Steckelmacher et al. Sample-efficient model-free reinforcement learning with off-policy critics. arXiv preprint arXiv:1903.04193, 2019.

[5] S. Tom et al. Prioritized experience replay. arXiv preprint arXiv:1511.05952, 2015.

[6] S. Fujimoto et al. Addressing Function Approximation Error in Actor-Critic Methods. International Conference on Machine Learning. 2018.

[7] J. Schrittwieser et al. Mastering atari, go, chess and shogi by planning with a learned model. arXiv preprint arXiv:1911.08265, 2019.

[8] R. K. Srivastava et al. Training Agents using Upside-Down Reinforcement Learning. arXiv preprint arXiv:1912.02877, 2019.

[9] J. Schmidhuber. Reinforcement Learning Upside Down: Don't Predict Rewards--Just Map Them to Actions. arXiv preprint arXiv:1912.02875, 2019.

About

This project contains various reinforcement/imitation learning experiments.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages