You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
@luke-conibear thanks for reporting this.
The patch-wise gradient accumulation, which is activated by use_patch_grad_acc =True, is an optimization that should only be used for training patched diffusion models.
We will push a fix that automatically disables it for all other types of models (e.g. non-patched diffusion and regression).
Version
Latest from main branch
On which installation method(s) does this occur?
Source
Describe the issue
Following this PR, the CorrDiff example has an error in the regression training:
In the PR, I see that the
use_patch_grad_acc
keyword argument was added toResidualLoss
but notRegressionLoss
.Should the same change be applied there?
The text was updated successfully, but these errors were encountered: