@inproceedings{4fd3726ff00e45bba3229c133a612f7d,
title = "Depth Dropout: Efficient Training of Residual Convolutional Neural Networks",
abstract = "Training state-of-the-art deep neural networks is computationally expensive and time consuming. In this paper we present a method that can reduce training time while at the same time maintain nearly the same accuracy as traditional training approaches. This allows for faster experimentation and better use of computational resource. Our method extends the well-known dropout technique by randomly removing entire network layers instead of individual neurons during training and hence reducing the number of expensive convolution operations needed per training iteration. We conduct experiments on object recognition using the CIFAR10 and ImageNet datasets to demonstrate the effectiveness of our approach. Our results show that we can train residual convolutional neural networks (ResNets) 17.5% faster with only 0.4% decrease in error rate or 34.1% faster with 1.3% increase in error rate compared to a baseline model. We also perform analysis on the trade-off between testing accuracy and training speedup as a function of the drop-out ratio.",
author = "Jian Guo and Stephen Gould",
note = "Publisher Copyright: {\textcopyright} 2016 IEEE.; 2016 International Conference on Digital Image Computing: Techniques and Applications, DICTA 2016 ; Conference date: 30-11-2016 Through 02-12-2016",
year = "2016",
month = dec,
day = "22",
doi = "10.1109/DICTA.2016.7797032",
language = "English",
series = "2016 International Conference on Digital Image Computing: Techniques and Applications, DICTA 2016",
publisher = "Institute of Electrical and Electronics Engineers Inc.",
editor = "Liew, {Alan Wee-Chung} and Jun Zhou and Yongsheng Gao and Zhiyong Wang and Clinton Fookes and Brian Lovell and Michael Blumenstein",
booktitle = "2016 International Conference on Digital Image Computing",
address = "United States",
}