Replies: 1 comment 1 reply
-
Hello @ghsanti Thank you for pointing the issue out. Can you please set I suspect division by 0 or near-0 numbers are the culprit here. Best Regards, |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
I've just read this issue in tf-keras, and ended up testing a simple convnet with mnist to understand the problem.
I suspect the reason may be due to BatchNormalization, but don't know enough to understand why.
Here is a simple image comparing the mean absolute error for
model(X, training=True)
andmodel.predict(X)
which disappears istraining=False
:But why would that make the outputs different, between
model(X, training=False)
(ormodel.predict
), andmodel(X, training=True)
I don't have a colab link, but the snippet I used is:
Is there a simple explanation? I should read the code but maybe someone knows already.
Anyone that could help, please?
Beta Was this translation helpful? Give feedback.
All reactions