A misclassification cost-minimizing evolutionary–neural classification approach
✍ Scribed by Parag Pendharkar; Sudhir Nanda
- Book ID
- 102543123
- Publisher
- John Wiley and Sons
- Year
- 2006
- Tongue
- English
- Weight
- 155 KB
- Volume
- 53
- Category
- Article
- ISSN
- 0894-069X
No coin nor oath required. For personal study only.
✦ Synopsis
Abstract
Machine learning algorithms that incorporate misclassification costs have recently received considerable attention. In this paper, we use the principles of evolution to develop and test an evolutionary/genetic algorithm (GA)‐based neural approach that incorporates asymmetric Type I and Type II error costs. Using simulated, real‐world medical and financial data sets, we compare the results of the proposed approach with other statistical, mathematical, and machine learning approaches, which include statistical linear discriminant analysis, back‐propagation artificial neural network, integrated cost preference‐based linear mathematical programming‐based minimize squared deviations, linear integrated cost preference‐based GA, decision trees (C 5.0, and CART), and inexpensive classification with expensive tests algorithm. Our results indicate that the proposed approach incorporating asymmetric error costs results in equal or lower holdout sample misclassification cost when compared with the other statistical, mathematical, and machine learning misclassification cost‐minimizing approaches. © 2006 Wiley Periodicals, Inc. Naval Research Logistics, 2006.
📜 SIMILAR VOLUMES
A new, dynamic, tree structured network, the Competitive Evolutionary Neural Tree (CENT) is introduced. The network is able to provide a hierarchical classification of unlabelled data sets. The main advantage that the CENT offers over other hierarchical competitive networks is its ability to self de
The task of classifying observations into known groups is a common problem in decision making. A wealth of statistical approaches, commencing with Fisher's linear discriminant function, and including variations to accommodate a variety of modeling assumptions, have been proposed. In addition, nonpar