You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I have the same question, since I've seen implementations that normalize these gradients. However, if batch size is always the same constant, then it doesn't matter that much and you can just adjust the learning rate, right?
This gradient for actor needs to be normalized over batch_size?
DDPG/actor_network_bn.py
Line 37 in 18825ee
The text was updated successfully, but these errors were encountered: