Skip to content

How to implement parallel training for model-based RL #292

Answered by Gaiejj
jintaoXue asked this question in Q&A
Discussion options

You must be logged in to vote

The implementation of parallelization is theoretically plausible.

  • A3C Parallel is established with the logic of the parallel code in the on-policy part, as indicated in the policy_gradient.py file, which contains the core of A3C. The crux lies in modifying steps_per_epoch based on the number of parallels, averaging the gradients of the actor-critic and world model before the optimizer's step.
  • Environment Parallel uses the vectorized environment for parallel data collection. You can also refer to omnsiafe/envs/safety_gymnasium.py to enable a vector environment.

Initial attempts at implementing parallelization were made, but subsequent tests suggested that the overhead of planning might o…

Replies: 1 comment 1 reply

Comment options

You must be logged in to vote
1 reply
@jintaoXue
Comment options

Answer selected by jintaoXue
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
2 participants