Advanced search
Start date
Betweenand


An immune and a gradient-based method to train multi-layer perceptron neural networks

Full text
Author(s):
Pasti, Rodrigo ; de Castro, Leandro Nunes ; IEEE
Total Authors: 3
Document type: Journal article
Source: 2007 IEEE INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, VOLS 1-6; v. N/A, p. 2-pg., 2006-01-01.
Abstract

Multi-layer perceptron (MLP) neural network training can be seen as a special case of function approximation, where no explicit model of the data is assumed. In its simplest form, it corresponds to finding an appropriate set of weights that minimize the network training and generalization errors. Various methods can be used to determine these weights, from standard optimization methods (e.g., gradient-based algorithms) to bio-inspired heuristics (e.g., evolutionary algorithms). Focusing on the problem of finding appropriate weight vectors for MLP networks, this paper proposes the use of an immune algorithm and a second-order gradient-based technique to train MLPs. Results are obtained for classification and function approximation tasks and the different approaches are compared in relation to the types of problems they are more suitable for. (AU)