Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

About the input of policy #9

Open
Liuxueyi opened this issue May 6, 2024 · 2 comments
Open

About the input of policy #9

Liuxueyi opened this issue May 6, 2024 · 2 comments

Comments

@Liuxueyi
Copy link

Liuxueyi commented May 6, 2024

In the paper, there is a state that

leverage the output state policy in nonmemory environments and the full state policy or hidden state policy within memory environments.

But in the configs.yaml, there is only in the POPGym task where the actor.inputs is set to be [stoch, hidden]. Why? Looking forward to your reply.

@artemZholus
Copy link
Collaborator

Hi @Liuxueyi .

the sentence you quoted is a general recommendation for working with new environments. There is another part of the paper where we state that in popgym we use the hidden state policy, so this config reflects that.

@Liuxueyi
Copy link
Author

Liuxueyi commented May 7, 2024

Thanks for your reply!
By the way, for the atari100k environment and pong task, what should be the input of policy(the default is [deter, stoch, hidden])? And in the issue #8 , when set the parameter run.script: train, there isn't any error. Is this feasible and will it affect the final performance of the algorithm?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants