You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
using active function called RuLU(max(0,x)),which is non-saturating,but is not zero-meaning.the initialization is crucial.
using multiple GPUs by separating net like pic
using local response normalizaton
using overlapping pool
Reducing Overfitting
Date Augmentation
horizontal reflections
random 224*224 patch from 256 *256 (not random (256-224)^2 * 2 = 2048)
altering the intensities of the TGB channels
Drpout
0.5 output set zero
Dropout roughly doubles the number of iterations required to converge.
train
using momentum of 0.9
VGGNet
VGGNet has muti-structure followed the pic
using smaller convolutional cell and more deeper than AlexNet
training image size is random sampling from the pic,which is rescaling S.
We consider two approaches for setting the training scale S. The first is to fix S, which correspondsto single-scale training (note that image content within the sampled crops can still represent multi-scale image statistics). In our experiments, we evaluated models trained at two fixed scales: S =256 (which has been widely used in the prior art (Krizhevsky et al., 2012; Zeiler & Fergus, 2013;Sermanet et al., 2014)) and S = 384. Given a ConvNet configuration, we first trained the networkusing S = 256. To speed-up training of the S = 384 network, it was initialised with the weightspre-trained with S = 256, and we used a smaller initial learning rate of 10−3.
The second approach to setting S is multi-scale training, where each training image is individuallyrescaled by randomly sampling S from a certain range [Smin,Smax] (we used Smin = 256 andSmax = 512). Since objects in images can be of different size, it is beneficial to take this into accountduring training. This can also be seen as training set augmentation by scale jittering, where a single model is trained to recognise objects over a wide range of scales. For speed reasons, we trainedmulti-scale models by fine-tuning all layers of a single-scale model with the same configuration,pre-trained with fixed S = 384.
The text was updated successfully, but these errors were encountered:
the summary of AlexNet and VGGNet
AlexNet VGGNet
AlexNet
AlexNet has five CNN and three full connection
like this
using active function called RuLU(max(0,x)),which is non-saturating,but is not zero-meaning.the initialization is crucial.
using multiple GPUs by separating net like pic
using local response normalizaton
using overlapping pool
Reducing Overfitting
Date Augmentation
Drpout
0.5 output set zero
Dropout roughly doubles the number of iterations required to converge.
train
using momentum of 0.9
VGGNet
VGGNet has muti-structure followed the pic
using smaller convolutional cell and more deeper than AlexNet
training image size is random sampling from the pic,which is rescaling S.
The text was updated successfully, but these errors were encountered: