Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Have you ever tried "Pendulum-V0"? #3

Open
yanpanlau opened this issue Sep 4, 2016 · 1 comment
Open

Have you ever tried "Pendulum-V0"? #3

yanpanlau opened this issue Sep 4, 2016 · 1 comment

Comments

@yanpanlau
Copy link

Thanks for the nice code. I am trying to re-produce the result in "Pendulum-V0" using a3c_cont.py but it seems the model fail to converge. I have tried various method like experience reply but still not working. It would be nice if you can test it out and we can discuss together Cheers.

@originholic
Copy link
Owner

Hi @yanpanlau , thanks for trying out the code. Unfortunately, I didn't actually test it with gym's Pendulum-v0, since this repo is very experimental for testing out my "batch" method.

If you are interested in getting the continuous actions to work, it is better to use other frameworks like miyosuda/async_deep_reinforce or coreylynch/async-rl. And just change the loss function for the policy and models as mentioned in the deepmind's async paper. Many thanks

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants