-
Notifications
You must be signed in to change notification settings - Fork 1.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Fine tuning losses #89
Comments
I don't think it is normal (I'm in the same boat). What I noticed is that if I improve FID from, say, 100 to 50, it jumps back to 90 or so upon resume. To be clear, FID will be 50 immediately on resume but then veer off to 90 in several ticks and only then will start to gradually subside. So I strongly suspect this behaviour is due to training schedule, Adam or both. In stylegan2-ada, you could start with kimg 10000 or whatever thus entering fine-tuning regime. If anyone knows how to tweak the training schedule, please share! Thanks |
Sorry, meant to say in StyleGAN2 rather than StyleGAN2-ada |
Ok nice find! I actually think the above has a good chance of fixing the issue. I suspect that the immediate divergence you're seeing is due to the augment strength being reset to 0 when resuming. The PR above seems to fix that. I'm going to give it a go and see where it takes me. |
I doubt that changing the augmentation strength is going to help in this case. But it is easy to test. |
Just as a heads up - playing around with #3, I seem to be getting much better results, whether it's because of augmentation strength or one of the other changes. |
Hello, Thank you for making this code base open-source, it's great!
I'm having the following issue: I'm fine-tuning the ffhq model on my own dataset. Since I'm training on colab, I have to do this piecewise, so I end up training as long as possible, then restarting from the latest snapshot.
The problem is that when I look at the losses, they seem to start from scratch every time. I includea screnshot of losses for two subsequent runs. I call train.py with the following arguments (other than the snapshot and data paths)
--aupipe=bg --gamma=10 --cfg=paper256 --mirror=1 --snap=10 --metrics=none
Is this normal would you say? What's then the best way of getting a sense of progress (other than manually inspecting outputs)? Thanks!
The text was updated successfully, but these errors were encountered: