Have a personal or library account? Click to login
Backpropagation generalized delta rule for the selective attention Sigma-if artificial neural network Cover

Backpropagation generalized delta rule for the selective attention Sigma-if artificial neural network

By: Maciej Huk  
Open Access
|Jun 2012

Abstract

In this paper the Sigma-if artificial neural network model is considered, which is a generalization of an MLP network with sigmoidal neurons. It was found to be a potentially universal tool for automatic creation of distributed classification and selective attention systems. To overcome the high nonlinearity of the aggregation function of Sigma-if neurons, the training process of the Sigma-if network combines an error backpropagation algorithm with the self-consistency paradigm widely used in physics. But for the same reason, the classical backpropagation delta rule for the MLP network cannot be used. The general equation for the backpropagation generalized delta rule for the Sigma-if neural network is derived and a selection of experimental results that confirm its usefulness are presented.

DOI: https://doi.org/10.2478/v10006-012-0034-5 | Journal eISSN: 2083-8492 | Journal ISSN: 1641-876X
Language: English
Page range: 449 - 459
Published on: Jun 28, 2012
In partnership with: Paradigm Publishing Services
Publication frequency: 4 issues per year

© 2012 Maciej Huk, published by University of Zielona Góra
This work is licensed under the Creative Commons License.

Volume 22 (2012): Issue 2 (June 2012)