Loading...

A complementary method for preventing hidden neurons' saturation in feed forward neural networks training

Moallem, P ; Sharif University of Technology | 2010

893 Viewed
  1. Type of Document: Article
  2. Publisher: 2010
  3. Abstract:
  4. In feed forward neural networks, hidden layer neurons' saturation conditions, which are the cause of flat spots on the error surface, is one of the main disadvantages of any conventional gradient descent learning algorithm. In this paper, we propose a novel complementary scheme for the learning based on a suitable combination of anti saturated hidden neurons learning process and accelerating methods like the momentum term and the parallel tangent technique. In our proposed method, a normalized saturation criterion (NSC) of hidden neurons, which is introduced in this paper, is monitored during learning process. When the NSC is higher than a specified threshold, it means that the algorithm moves towards a flat spot as the hidden neurons fall into saturation condition. In this case, in order to suppress the saturation of hidden neurons, a conventional gradient descent learning method can be accompanied by the proposed complementary gradient descent saturation prevention scheme. When the NSC assumes small values, no saturation detected and the network operates in its normal condition. Therefore, application of a saturation prevention scheme is not recommended. We have evaluated the proposed complementary method in accompaniment to the gradient descent plus momentum and parallel tangent, two conventional improvements on learning methods. We have recorded remarkable improvements in convergence success as well as generalization in some well known benchmarks
  5. Keywords:
  6. Complementary methods ; Error surface ; Flat spot ; Gradient descent ; Gradient descent learning algorithm ; Hidden layer neurons ; Hidden neurons ; Hidden neurons' saturation ; Learning methods ; Learning process ; Momentum term ; Normal condition ; Normalized saturation criterion ; Parallel tangent gradient ; Saturation conditions ; Backpropagation ; Learning algorithms ; Learning systems ; Network layers ; Topography ; Neural networks
  7. Source: Iranian Journal of Electrical and Computer Engineering ; Volume 9, Issue 2 , SUMMER-FALL , 2010 , Pages 127-133 ; 16820053 (ISSN)
  8. URL: http://en.journals.sid.ir/ViewPaper.aspx?ID=190099