Skip to content

mbhenaff/neural-e3

Repository files navigation

Explicit Explore-Exploit Algorithms in Continuous State Spaces

Code for the Neural-E3 algorithm described in the paper Explicit Explore-Exploit Algorithms in Continuous State Spaces. All code was written with Python 3.6 and PyTorch v1.1.0.

To run Neural-E3 on the combination lock tasks, do:

cd combolock/
python Experiment.py --alg neural-e3 --horizon 5 --lr 0.01 --n_playouts 200 --n_ensemble 5 --seed 1 --env_param_1 0.1 

To run Neural-E3 on continuous control tasks, do:

python train_e3.py -env {mountaincar, acrobot} -n_exploration_epochs 10 -n_trajectories 10 -eps 0.1 -n_model_updates 200 -n_exploration_epochs 10 

To run it on maze tasks, do:

cd src/
python train_e3.py -env {maze5, maze10, maze20} -n_exploration_epochs 10  -lambda_r 0.01

The code for running the baseline algorithms (DQN, PPO and PPO+RND) is in the 'baselines' subfolder. The DQN uses the OpenAI Baselines repo and the PPO+RND code is built on the DeepRL Pytorch repo. The combination lock environment is built on the codebase of Du et al..

Releases

No releases published

Packages

No packages published

Languages