# Nearest neighbor search

## Optimization problem in computer science / From Wikipedia, the free encyclopedia

#### Dear Wikiwand AI, let's keep it short by simply answering these key questions:

Can you list the top facts and stats about Nearest neighbor search?

Summarize this article for a 10 year old

**Nearest neighbor search** (**NNS**), as a form of **proximity search**, is the optimization problem of finding the point in a given set that is closest (or most similar) to a given point. Closeness is typically expressed in terms of a dissimilarity function: the less similar the objects, the larger the function values.

Formally, the nearest-neighbor (NN) search problem is defined as follows: given a set *S* of points in a space *M* and a query point *q* ā *M*, find the closest point in *S* to *q*. Donald Knuth in vol. 3 of *The Art of Computer Programming* (1973) called it the **post-office problem**, referring to an application of assigning to a residence the nearest post office. A direct generalization of this problem is a *k*-NN search, where we need to find the *k* closest points.

Most commonly *M* is a metric space and dissimilarity is expressed as a distance metric, which is symmetric and satisfies the triangle inequality. Even more common, *M* is taken to be the *d*-dimensional vector space where dissimilarity is measured using the Euclidean distance, Manhattan distance or other distance metric. However, the dissimilarity function can be arbitrary. One example is asymmetric Bregman divergence, for which the triangle inequality does not hold.[1]

Oops something went wrong: