File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Conference Paper: On the importance of network architecture in training very deep neural networks

TitleOn the importance of network architecture in training very deep neural networks
Authors
Issue Date2016
Citation
ICSPCC 2016 - IEEE International Conference on Signal Processing, Communications and Computing, Conference Proceedings, 2016, article no. 7753635 How to Cite?
AbstractVery deep neural networks with hundreds or more layers have achieved significant success in a variety of vision tasks spanning from image classification, detection, to image captioning. However, simply stacking more layers in the convolution operation could suffer from the gradient vanishing problem and thus could not lower down the training loss further. The residual network [1] pushes the model's depth to extremely deep by proposing an identity mapping plus a residual learning term and addresses the gradient back-propagation bottleneck well. In this paper, we investigate the residual module in great extent by analyzing the structure ordering of different blocks and modify them one by one to achieve lower test error on CIFAR-10 dataset. One key observation is that removing the original ReLU activation could facilitate the gradient propagation in the identity mapping path. Moreover, inspired by the ResNet block, we propose a random-jump scheme to skip some residual connections during training, i.e., lower features could jump to any subsequent layers and bypass its transformations directly to the higher level. Such an upgrade to the network structure not only saves training time but also obtains better performance.
Persistent Identifierhttp://hdl.handle.net/10722/351372

 

DC FieldValueLanguage
dc.contributor.authorChi, Zhizhen-
dc.contributor.authorLi, Hongyang-
dc.contributor.authorWang, Jingjing-
dc.contributor.authorLu, Huchuan-
dc.date.accessioned2024-11-20T03:55:53Z-
dc.date.available2024-11-20T03:55:53Z-
dc.date.issued2016-
dc.identifier.citationICSPCC 2016 - IEEE International Conference on Signal Processing, Communications and Computing, Conference Proceedings, 2016, article no. 7753635-
dc.identifier.urihttp://hdl.handle.net/10722/351372-
dc.description.abstractVery deep neural networks with hundreds or more layers have achieved significant success in a variety of vision tasks spanning from image classification, detection, to image captioning. However, simply stacking more layers in the convolution operation could suffer from the gradient vanishing problem and thus could not lower down the training loss further. The residual network [1] pushes the model's depth to extremely deep by proposing an identity mapping plus a residual learning term and addresses the gradient back-propagation bottleneck well. In this paper, we investigate the residual module in great extent by analyzing the structure ordering of different blocks and modify them one by one to achieve lower test error on CIFAR-10 dataset. One key observation is that removing the original ReLU activation could facilitate the gradient propagation in the identity mapping path. Moreover, inspired by the ResNet block, we propose a random-jump scheme to skip some residual connections during training, i.e., lower features could jump to any subsequent layers and bypass its transformations directly to the higher level. Such an upgrade to the network structure not only saves training time but also obtains better performance.-
dc.languageeng-
dc.relation.ispartofICSPCC 2016 - IEEE International Conference on Signal Processing, Communications and Computing, Conference Proceedings-
dc.titleOn the importance of network architecture in training very deep neural networks-
dc.typeConference_Paper-
dc.description.naturelink_to_subscribed_fulltext-
dc.identifier.doi10.1109/ICSPCC.2016.7753635-
dc.identifier.scopuseid_2-s2.0-85006915140-
dc.identifier.spagearticle no. 7753635-
dc.identifier.epagearticle no. 7753635-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats