Skip to content

4kasha/CartPole_PPO

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

4 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

CartPole via PPO - PyTorch implementation

Description

You will train an agent in CartPole-v0 (OpenAI Gym) environment via Proximal Policy Optimization (PPO) algorithm with GAE.

A reward of +1 is provided for every step taken, and a reward of 0 is provided at the termination step. The state space has 4 dimensions and contains the cart position, velocity, pole angle and pole velocity at tip. Given this information, the agent has to learn how to select best actions. Two discrete actions are available, corresponding to:

  • 0 - 'Push cart to the left'
  • 1 - 'Push cart to the right'

For more details, see the wiki.

For training results and making animation, see train.ipynb.

Dependencies

  • Python 3.6
  • PyTorch 0.4.0
  • OpenAI Gym 0.10.5 (for Installation, see this.)

References

About

CartPole-v0 via PPO with GAE, PyTorch

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published