Article ID Journal Published Year Pages File Type
407897 Neurocomputing 2013 16 Pages PDF
Abstract

Most methods for the evolutionary generation of multi-layer perceptron classifiers use a divide-and-conquer strategy, where the tasks of feature selection, structure design, and weight training are performed separately. The concurrent evolution of the whole classifier has been seldom attempted and its effectiveness has never been exhaustively benchmarked. This paper presents an experimental study on the merits of this latter approach. Two schemes were investigated. The first method evolves simultaneously the neural network structure and input feature vector, and trains via a standard learning procedure the candidate solutions (wrapper approach). The second method evolves simultaneously the whole classifier (embedded approach). The performance of these two algorithms was compared to that of two manual and two automatic neural network optimisation techniques on thirteen well-known pattern recognition benchmarks. The experimental results revealed the specific strengths and weaknesses of the six algorithms. Overall, the evolutionary embedded method obtained good results in terms of classification accuracy and compactness of the solutions. The tests indicated that the outcome of the feature selection task has a major impact on the accuracy and compactness of the solutions. Evolutionary algorithms perform best on feature spaces of small and medium size, and were the most effective at rejecting redundant features. Classical filter-based algorithms based on feature correlation are preferable on undersampled data sets. Correlation- and saliency-based selection was the most effective method in the presence of a large number of irrelevant features. The applicability and performance of the wrapper algorithm was severely limited by the computational costs of the approach.

► The evolution of ANN classifiers was benchmarked against classical and manual methods. ► The best method is to evolve concurrently data features, ANN structure and weights. ► The quality of the solutions is mostly affected by the feature selection task. ► EAs excel when the feature set is of moderate size and contains redundant instances. ► Correlation-based selection excels on large feature sets containing irrelevant elements.

Related Topics
Physical Sciences and Engineering Computer Science Artificial Intelligence
Authors
,