-
Notifications
You must be signed in to change notification settings - Fork 1.7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Feature Request] Can PPO support graph style spaces? #1280
Comments
Related: #219 (comment)
Would you volunteer to add support for it?
After a quick github search, there is actually some code online: https://github.com/AndrewPaulChester/oracle-sage/blob/3d5898cff22c2414d8199fe9541b67e503e29804/sage/agent/graph_feedback_policy.py#L146 |
Thanks for your information, it's very helpful to me. I also found another repo https://github.com/YinqiangZhang/custom_stable_baselines that is very close to what I need, so I may directly use it. |
@araffin: One big hurdle IMO can be removed by allowing action and value modules directly output the actions, values respectively instead of only outputting the embeddings. Thus removing the need for 'extra linear layer' mentioned in docs which might not be the right thing to do in this usecase. |
My approach to tackle:
Would it be a good idea to create |
🚀 Feature
Support graph style data structure as the observation and action spaces for RL algorithms like PPO or others.
Motivation
After version 0.25.0, gym has support graph style observation or action spaces. Some remarkable works like A graph placement methodology for fast chip design has proved that using PPO combined with GNN feature extractor can reach an excellent result. Since GNN has become a common neural network architecture, it should be supported for the environment spaces.
Pitch
No response
Alternatives
No response
Additional context
No response
Checklist
The text was updated successfully, but these errors were encountered: