Rectified activation units (rectifiers) are essential for state-of-the-art
neural networks. In this work, we study rectifier neural networks for image
classification from two aspects. First, we propose a Parametric Rectified
Linear Unit (PReLU) that generalizes the traditional rectified unit. PReLU
improves model fitting with nearly zero extra computational cost and little
overfitting risk. Second, we derive a robust initialization method that
particularly considers the rectifier nonlinearities. This method enables us to
train extremely deep rectified models directly from scratch and to investigate
deeper or wider network architectures. Based on our PReLU networks
(PReLU-nets), we achieve 4.94\% top-5 test error on the ImageNet 2012
classification dataset. This is a 26\% relative improvement over the ILSVRC 2014
winner (GoogLeNet, 6.66\%). To our knowledge, our result is the first to surpass
human-level performance (5.1\%, Russakovsky et al.) on this visual recognition
challenge.
%0 Generic
%1 he2015delving
%A He, Kaiming
%A Zhang, Xiangyu
%A Ren, Shaoqing
%A Sun, Jian
%D 2015
%K imported
%T Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification
%U http://arxiv.org/abs/1502.01852
%X Rectified activation units (rectifiers) are essential for state-of-the-art
neural networks. In this work, we study rectifier neural networks for image
classification from two aspects. First, we propose a Parametric Rectified
Linear Unit (PReLU) that generalizes the traditional rectified unit. PReLU
improves model fitting with nearly zero extra computational cost and little
overfitting risk. Second, we derive a robust initialization method that
particularly considers the rectifier nonlinearities. This method enables us to
train extremely deep rectified models directly from scratch and to investigate
deeper or wider network architectures. Based on our PReLU networks
(PReLU-nets), we achieve 4.94\% top-5 test error on the ImageNet 2012
classification dataset. This is a 26\% relative improvement over the ILSVRC 2014
winner (GoogLeNet, 6.66\%). To our knowledge, our result is the first to surpass
human-level performance (5.1\%, Russakovsky et al.) on this visual recognition
challenge.
@misc{he2015delving,
abstract = {{Rectified activation units (rectifiers) are essential for state-of-the-art
neural networks. In this work, we study rectifier neural networks for image
classification from two aspects. First, we propose a Parametric Rectified
Linear Unit (PReLU) that generalizes the traditional rectified unit. PReLU
improves model fitting with nearly zero extra computational cost and little
overfitting risk. Second, we derive a robust initialization method that
particularly considers the rectifier nonlinearities. This method enables us to
train extremely deep rectified models directly from scratch and to investigate
deeper or wider network architectures. Based on our PReLU networks
(PReLU-nets), we achieve 4.94\% top-5 test error on the ImageNet 2012
classification dataset. This is a 26\% relative improvement over the ILSVRC 2014
winner (GoogLeNet, 6.66\%). To our knowledge, our result is the first to surpass
human-level performance (5.1\%, Russakovsky et al.) on this visual recognition
challenge.}},
added-at = {2017-07-19T15:29:59.000+0200},
archiveprefix = {arXiv},
author = {He, Kaiming and Zhang, Xiangyu and Ren, Shaoqing and Sun, Jian},
biburl = {https://www.bibsonomy.org/bibtex/2dd7f625aa57c0d1b9f91f8be8135a513/andreashdez},
citeulike-article-id = {13515098},
citeulike-linkout-0 = {http://arxiv.org/abs/1502.01852},
citeulike-linkout-1 = {http://arxiv.org/pdf/1502.01852},
day = 6,
eprint = {1502.01852},
interhash = {0b13164e63812ff4bafd89b6139fe961},
intrahash = {dd7f625aa57c0d1b9f91f8be8135a513},
keywords = {imported},
month = feb,
posted-at = {2016-05-01 20:47:22},
priority = {2},
timestamp = {2017-07-19T15:31:02.000+0200},
title = {{Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification}},
url = {http://arxiv.org/abs/1502.01852},
year = 2015
}