Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

RewardScaler #4

Open
NikEyX opened this issue Jul 15, 2019 · 0 comments
Open

RewardScaler #4

NikEyX opened this issue Jul 15, 2019 · 0 comments

Comments

@NikEyX
Copy link

NikEyX commented Jul 15, 2019

I noticed the below:

class RewardScaler(gym.RewardWrapper):
    """
    Bring rewards to a reasonable scale for PPO. This is incredibly important
    and effects performance a lot.
    """
    def reward(self, reward):
        return reward * 0.01

Is there a good explanation for this? I would have thought it should converge regardless? Why does it affect convergence speed so much? How did you even find this to be a problem? How can I know whether my rewards (on a different game) are too large or need to be scaled, are there good indicators for that besides performance (which could be affected by any parameters)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant