Abstract
This work presents a class of functions serving as generalized neuron
models to be used in artificial neural networks. They are cast into the
common framework of computing a similarity function, a flexible
definition of a neuron as a pattern recognizer. The similarity endows
the model with a clear conceptual view and serves as a unification
cover for many of the existing neural models, including those
classically used for the MultiLayer Perceptron (MLP) and most of those
used in Radial Basis Function
Networks (RBF). These families of models are conceptually unified and
their relation is clarified.
The possibilities of deriving new instances are explored and several
neuron models --representative of their families-- are proposed.
The similarity view naturally leads to further extensions of the models
to handle heterogeneous information, that is to say, information coming
from sources radically different in character, including continuous and
discrete (ordinal) numerical quantities, nominal (categorical)
quantities, and fuzzy quantities. Missing data are also explicitly
considered. A neuron of this class is called an heterogeneous neuron
and any neural structure making use of them is an Heterogeneous Neural
Network (HNN), regardless of the specific architecture or learning
algorithm. Among them, in this work we concentrate on feed-forward
networks, as the initial focus of study. The learning procedures may
include a great variety of techniques, basically divided in
derivative-based methods (such as the conjugate gradient)and
evolutionary ones (such as variants of genetic algorithms).
In this Thesis we also explore a number of directions towards the
construction of better neuron models --within an integrant envelope--
more adapted to the problems they are meant to solve.
It is described how a certain generic class of heterogeneous models
leads to a satisfactory performance, comparable, and often better, to
that of classical neural models, especially in the presence of
heterogeneous information, imprecise or incomplete data, in a wide
range of domains, most of them corresponding to real-world problems.