Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Why retrain at about 95 epoch with data aug disabled #47

Open
DeppMeng opened this issue Jan 30, 2018 · 3 comments
Open

Why retrain at about 95 epoch with data aug disabled #47

DeppMeng opened this issue Jan 30, 2018 · 3 comments

Comments

@DeppMeng
Copy link

Hi, I am trying to reproduce the resnet imagenet result on mxnet.

I noticed that you have successfully reproduced the resnet results from facebook,
but you mentioned that

when epoch is about 95, cancel the scale/color/aspect augmentation during training, this can be done by only comment out 6 lines of the code, like this:

I didn't quite understand why we need this procedure to get a similar results because I as far as I know Torch/PyTorch don't need this procedure.
Is that means if we don't apply this procedure, the result will be suffered (lower than the paper claimed)?

Thanks

@jamesdeep
Copy link

I have the same question, any clus till now?

Hi, I am trying to reproduce the resnet imagenet result on mxnet.

I noticed that you have successfully reproduced the resnet results from facebook,
but you mentioned that

when epoch is about 95, cancel the scale/color/aspect augmentation during training, this can be done by only comment out 6 lines of the code, like this:

I didn't quite understand why we need this procedure to get a similar results because I as far as I know Torch/PyTorch don't need this procedure.
Is that means if we don't apply this procedure, the result will be suffered (lower than the paper claimed)?

Thanks

I have the same question, any clus till now?

@ky-du
Copy link

ky-du commented Dec 10, 2019

@DeppMeng I think data augmentation may change the overall distribution of the dataset. When trained with small learning rate, the network may learn such distorted distribution.

By the way, do you have the training log / script on Pytorch? Thanks :)

@DeppMeng
Copy link
Author

Sorry I am not doing research in this field now:)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants