Dropout technique

Dropout is one of the simplest and the most powerful regularization techniques. It prevents units from complex co-adapting by randomly dropping units from the network. [N. Srivastava et al., 2014] Below table is a list of famous deep networks which use dropout techniques.

Status of dropout technique usage in famous deep networks

Model Dropout layers Remark
AlexNet [Alex Krizhevsky et al., 2012] Used in two fully-connected layers Won the 2012 ILSVRC (ImageNet Large-Scale Visual Recognition Challenge)
ZFNet [Matthew D. Zeiler et al., 2013] Used in two fully-connected layers Won the 2013 ILSVRC
VGG Net [Karen Simonyan et al., 2014] Used in two fully-connected layers Best utilized with simple and deep CNN
GoogLeNet [Christian Szegedy et al., 2015] Used in one fully-connected layer Won the 2014 ILSVRC
Generative Adversarial Networks [Ian J. Goodfellow et al., 2014] Applied in training the discriminator net Various usage such as feature extraction, generating artificial images
Generating Image Descriptions [Adrej Karpathy et al., 2014] Used in all layers except in the recurrent layers Combination of CNNs and RNNs
Spatial Transformer Networks [Max Jaderberg et al., 2015] Used in all layers except the first convolutional layer Introduce of a Spatial Transformer module

Notices

The list will keep updated