Importance of initialization of weight matrices in deep learning neural networks

Authors

  • Nicholas Christopher A. Colina National Institute of Physics, University of the Philippines Diliman
  • Carlos E. Perez ceperez@alluviate.com
  • Francis N. C. Paraan National Institute of Physics, University of the Philippines Diliman

Abstract

The success of deep neural networks relies on optimized weight matrices are initialized in different ways. This work reports learning improvement in a six-layer deep neural network that is initialized with orthogonal weight matrices when compared to other commonly-used initialization schemes. An analysis of the eigenvalue spectra of the optimized solutions implies that the space of orthogonal weight matrices lies close to the manifold of learned states.

Downloads

Issue

Article ID

SPP-2016-PA-21

Section

Poster Session PA

Published

2016-08-18

How to Cite

[1]
NCA Colina, CE Perez, and FNC Paraan, Importance of initialization of weight matrices in deep learning neural networks, Proceedings of the Samahang Pisika ng Pilipinas 34, SPP-2016-PA-21 (2016). URL: https://proceedings.spp-online.org/article/view/SPP-2016-PA-21.