Improving the model convergence properties of classifier feed-forward MLP neural networks

A. R. Várkonyi-Kóczy, B. Tusor, J. Bukor

Research output: Contribution to journalArticle

Abstract

Recently, the application of Artificial Neural Networks (ANNs) has become very popular. Their success is due to the fact that they are able to learn complex input-output mappings and are able to find relationships in unstructured data sets. Further, neural nets are relatively easy to implement in any application. In the last years, classification has become one of the most significant research and application area of ANNs because these networks have proved to be very efficient in the field. Unfortunately, a big difficulty of the usage of feed-forward multilayer perceptron (MLP) neural nets with supervised learning is that in case of higher problem complexity, the NN model may not converge during the training or in better cases needs a long training time which scales with the structural parameters of the networks and the quantity of input data. However, the training can be done off-line, this disadvantage may limit the usage of NN models because the training has a non-negligible cost and further, can cause a possibly non-tolerable delay in the operation. In this chapter, to overcome these problems, a new training algorithm is proposed which in many cases is able to improve the convergence properties of NN models in complex real world classification problems. On one hand, the accuracy of the models can be increased while on the other hand the training time can be decreased. The new training method is based on the wellknown back-propagation algorithms, however with a significant difference: instead of the original input data, a reduced data set is used during the teaching phase. The reduction is the result of a complexity optimized classification procedure. In the resulted new, reduced input data set, each input sample is replaced by the center of the cluster to which it belongs and these cluster centers are used during the training (each element once). As result, new, complex ambiguous classification problems can be solved with acceptable cost and accuracy by using feed-forward MLP NNs.

Original languageEnglish
Pages (from-to)281-293
Number of pages13
JournalStudies in Fuzziness and Soft Computing
Volume317
DOIs
Publication statusPublished - Jan 1 2014

ASJC Scopus subject areas

  • Computer Science (miscellaneous)
  • Computational Mathematics

Fingerprint Dive into the research topics of 'Improving the model convergence properties of classifier feed-forward MLP neural networks'. Together they form a unique fingerprint.

  • Cite this