Article ID Journal Published Year Pages File Type
409811 Neurocomputing 2015 11 Pages PDF
Abstract

•It finds the optimal number of neighbours K for probabilistic KNN (PKNN).•It reconstructs the density of K for the PKNN.•It incorporates the uncertainty in K into the decision making.•It provides improved PKNN classification using Bayesian model averaging.

Probabilistic K-nearest neighbour (PKNN) classification has been introduced to improve the performance of the original K-nearest neighbour (KNN) classification algorithm by explicitly modelling uncertainty in the classification of each feature vector. However, an issue common to both KNN and PKNN is to select the optimal number of neighbours, K. The contribution of this paper is to incorporate the uncertainty in K into the decision making, and consequently to provide improved classification with Bayesian model averaging. Indeed the problem of assessing the uncertainty in K can be viewed as one of statistical model selection which is one of the most important technical issues in the statistics and machine learning domain. In this paper, we develop a new functional approximation algorithm to reconstruct the density of the model (order) without relying on time consuming Monte Carlo simulations. In addition, the algorithms avoid cross validation by adopting Bayesian framework. The performance of the proposed approaches is evaluated on several real experimental datasets.

Related Topics
Physical Sciences and Engineering Computer Science Artificial Intelligence
Authors
, ,