-
Notifications
You must be signed in to change notification settings - Fork 14
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Why I can't converge the network when I use your default setting(terminal: python run_atari_dqn.py)? #6
Comments
You are running PongNoFrameskip-v4? and for how long did you let it train? I have another repository for how to do the plottings: https://github.com/BY571/Tensorboard-Plotting |
I checked yesterday and noticed the same. for CartPole-v0 and LunarLander-v2 it converges nicely. But as you noticed for Pong it was struggling. I'll have a look later might be something with the wrappers of the environment. Once I found out whats the problem I come back to you :) |
Em, Ok , Thanks!Let me try the CartPole-v0 and LunarLander-v2 environment firstly, and then try others. thanks for your answer :) |
how did your runs work out for cartpole and lunarlander? sry i was quite busy and couldn't dig deeper in the problem yet |
Hello, first of all thanks for sharing your amazing job. I want to know Why I can't converge the network when I use your default setting(terminal: python run_atari_dqn.py)? the average 100 reward value always in about -20.9, which can't increase as the time-step increase. And Can you share your code about how to draw good pictures like you folder ./imgs/ pictures. Thank your very much!!
The text was updated successfully, but these errors were encountered: