Improving gradient-based learning algorithms for large scale feedforward networks

M. Ventresca, H. R. Tizhoosh

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract

Large scale neural networks have many hundreds or thousands of parameters (weights and biases) to learn, and as a result tend to have very long training times. Small scale networks can be trained quickly by using second-order information, but these fail for large architectures due to high computational cost. Other approaches employ local search strategies, which also add to the computational cost. In this paper we present a simple method, based on opposite transfer functions which greatly improve the convergence rate and accuracy of gradientbased learning algorithms. We use two variants of the backpropagation algorithm and common benchmark data to highlight the improvements. We find statistically significant improvements in both converegence speed and accuracy.

Original languageEnglish (US)
Title of host publication2009 International Joint Conference on Neural Networks, IJCNN 2009
Pages3212-3219
Number of pages8
DOIs
StatePublished - 2009
Event2009 International Joint Conference on Neural Networks, IJCNN 2009 - Atlanta, GA, United States
Duration: Jun 14 2009Jun 19 2009

Publication series

NameProceedings of the International Joint Conference on Neural Networks

Conference

Conference2009 International Joint Conference on Neural Networks, IJCNN 2009
Country/TerritoryUnited States
CityAtlanta, GA
Period6/14/096/19/09

Keywords

  • Backpropgation
  • Gradient-based learning
  • Large scale networks
  • Opposite transfer functions
  • Opposition-based computing

ASJC Scopus subject areas

  • Software
  • Artificial Intelligence

Fingerprint

Dive into the research topics of 'Improving gradient-based learning algorithms for large scale feedforward networks'. Together they form a unique fingerprint.

Cite this