Éditeur(s) :
HAL CCSD IEEE Résumé : Object classification is a challenging task in computer vision. Many approaches have been proposed to extract meaningful descriptors from images and classifying them in a supervised learning framework. In this paper, we revisit the classic k-nearest neighbors (k-NN) classification rule, which has shown to be very effective when dealing with local image descriptors. However, k-NN still features some major drawbacks, mainly due to the uniform voting among the nearest prototypes in the feature space. In this paper, we propose a generalization of the classic k-NN rule in a supervised learning (boosting) framework. Namely, we redefine the voting rule as a strong classifier that linearly combines predictions from the k closest prototypes. To induce this classifier, we propose a novel learning algorithm, MLNN (Multiclass Leveraged Nearest Neighbors), which gives a simple procedure for performing prototype selection very efficiently. We tested our method on 12 categories of objects, and observed significant improvement over classic k-NN in terms of classification performances.
International Workshop on Image Analysis for Multimedia Interactive Services
Desenzano del Garda, Italy
hal-00481725
https://hal.archives-ouvertes.fr/hal-00481725 https://hal.archives-ouvertes.fr/hal-00481725/document https://hal.archives-ouvertes.fr/hal-00481725/file/pbnn_wiamis_10.pdf