k-nearest neighbor (k-NN) classiรฟcation is a well-known decision rule that is widely used in pattern classiรฟcation. However, the traditional implementation of this method is computationally expensive. In this paper we develop two e ective techniques, namely, template condensing and preprocessing, to
Nearest-neighbor classification with categorical variables
โ Scribed by Samuel E. Buttrey
- Publisher
- Elsevier Science
- Year
- 1998
- Tongue
- English
- Weight
- 736 KB
- Volume
- 28
- Category
- Article
- ISSN
- 0167-9473
No coin nor oath required. For personal study only.
โฆ Synopsis
A technique is presented for adopting nearest-neighbor classification to the case of categorical variables. The set of categories is mapped onto the real line in such a way as to maximize the ratio of total sum of squares to within-class sum of squares, aggregated over classes. The resulting real values then replace the categories, and nearest-neighbor classification proceeds with the Euclidean metric on these new values. Continuous variables can be included in this scheme with little added efort. This approach has been implemented in a computer program and tried on a number of data sets, with encouraging results.
Nearest-neighbor classification is a well-known and efective classification technique. With this scheme, an unknown item's distances to all known items are measured, and the unknown class is estimated by the class of the nearest neighbor or by the class most often represented among a set of nearest neighbors. This has proven effective in many examples, but an appropriate distance normalization is required when variables are scaled differently. For categorical variables "distance" is not even defined. In this paper categorical data values are replaced by real numbers in an optimal way: then those real numbers are used in nearest-neighbor classification. (~) 1998 Elsevier Science B.V. All rights reserved.
๐ SIMILAR VOLUMES
In this paper very simple nonparametric c l d i c a t i o n rule for mixtures of discrete and oonhuons random variables is described. It ie based on the method of neatest neighbor proposed by COVEB and HABT (1967). The bounds on the limit of the near& neighbor ruleriske are given. Both lower and upp
A novel neuralnet-based method of constructing optimized prototypes for nearest-neighbor classiรฟers is proposed. Based on an e ective classiรฟcation oriented error function containing class classiรฟcation and class separation components, the corresponding prototype and feature weight update rules are