(2015. 2) He Initialization
Submitted on 2015. 2
Kaiming He, Xiangyu Zhang, Shaoqing Ren and Jian Sun
Simple Summary
propose a Parametric Rectified Linear Unit (PReLU) that generalizes the traditional rectified unit. PReLU improves model fitting with nearly zero extra computational cost and little overfitting risk. Second, we derive a robust initialization method that particularly considers the rectifier nonlinearities. This method enables us to train extremely deep rectified models directly from scratch and to investigate deeper or wider network architectures.
PReLU, adaptively learns the parameters
The idea is that the ReLU will completely eliminate the negative signals and double the dispersion that they have to maintain.
achieves 4.94% top-5 error
Last updated