-
Notifications
You must be signed in to change notification settings - Fork 464
MultiDiscrete Gym Environments #176
Comments
I agree it makes sense to add support for
|
Both options look good to me. I'm concerned about how to map the multiple sets of discrete actions to the neural network outputs. If this get supported cleanly, I don't mind manually discretizing the continuous action spaces. |
I agree that there is another change that that this points to. The observation space can already be defined as a dictionary of multiple spaces. It would be nice to have something similar for action spaces. As it stands, gym environments only support |
Yes, completely agree. |
@bbalaji-ucsd check out the BoxDiscretization Action Filter. There's also a sample CARLA preset using it. Is this good enough for your purpose? |
@galnov very cool, i didn't know about this, and it does address the case i raised where an environment has a continuous action space that you want discretized. However, the original issue still stands which is that gym environments can define |
@galnov in gym, We could support this in coach with |
@galnov Is the acceptance criteria for this issue an agent that implements multi dimensional discrete RL? |
Gym API supports MultiDiscrete action spaces:
https://github.com/openai/gym/blob/master/gym/spaces/multi_discrete.py
This is useful when you want to discretize a continuous control problem, a technique common in literature: https://arxiv.org/abs/1808.00177
But MultiDiscrete action spaces are ignored in Coach:
https://github.com/NervanaSystems/coach/blob/master/rl_coach/environments/gym_environment.py#L367
Can you please add support for them?
As a concrete use case, I would like to (independently) discretize the steering and throttle actions in DeepRacer:
https://github.com/awslabs/amazon-sagemaker-examples/blob/master/reinforcement_learning/rl_deepracer_robomaker_coach_gazebo/src/robomaker/environments/deepracer_env.py#L320
The text was updated successfully, but these errors were encountered: