BDBComp
Parceria:
SBC
Global Optimization Methods for Designing and Training Neural Networks

Akio YamazakiTeresa Bernarda LudermirMarcílio Carlos Pereira de Souto

This paper shows results of two approaches for the optimization of neural networks: one uses simulated annealing for optimizing both architectures and weights combined with backpropagation for fine tuning, while the other uses tabu search for the same purpose. Both approaches generate networks with good generalization performance (mean classification error of 1.68% for simulated annealing and 0.64% for tabu search) and low complexity (mean number of connections of 11.15 out of 36 for simulated annealing and 11.62 out of 36 for tabu search) for an odor recognition task in an artificial nose.

http://csdl.computer.org/comp/proceedings/sbrn/2002/1709/00/17090136abs.htm

Caso o link acima esteja inválido, faça uma busca pelo texto completo na Web: Buscar na Web

Biblioteca Digital Brasileira de Computação - Contato: bdbcomp@lbd.dcc.ufmg.br
     Mantida por:
LBD