SOTAVerified

A Theory-Based Evaluation of Nearest Neighbor Models Put Into Practice

2018-10-11NeurIPS 2018Unverified0· sign in to hype

Hendrik Fichtenberger, Dennis Rohde

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

In the k-nearest neighborhood model (k-NN), we are given a set of points P, and we shall answer queries q by returning the k nearest neighbors of q in P according to some metric. This concept is crucial in many areas of data analysis and data processing, e.g., computer vision, document retrieval and machine learning. Many k-NN algorithms have been published and implemented, but often the relation between parameters and accuracy of the computed k-NN is not explicit. We study property testing of k-NN graphs in theory and evaluate it empirically: given a point set P R^ and a directed graph G=(P,E), is G a k-NN graph, i.e., every point p P has outgoing edges to its k nearest neighbors, or is it -far from being a k-NN graph? Here, -far means that one has to change more than an -fraction of the edges in order to make G a k-NN graph. We develop a randomized algorithm with one-sided error that decides this question, i.e., a property tester for the k-NN property, with complexity O(n k^2 / ^2) measured in terms of the number of vertices and edges it inspects, and we prove a lower bound of (n / k). We evaluate our tester empirically on the k-NN models computed by various algorithms and show that it can be used to detect k-NN models with bad accuracy in significantly less time than the building time of the k-NN model.

Tasks

Reproductions