High-Dimensional Computational Geometry 115
The Hausdoff metric contains a copy of
∞
, and hence, it is a natural host space. A few
such embeddings into
∞
, for different types of metrics X isshownin[47].
Are there even better host spaces?
∞
is another natural target, especially that it contains
any other metric. However, as a host,
∞
often requires prohibitively high dimension: even
embedding the Hamming cube {0, 1}
d
into
∞
requires dimension exponential in d (see
arguments from [67, Chapter 1, Section 8]). Hence, new candidates are needed.
It turns out that the mixture of norms is a qualitatively better host. Let us define what
we mean by mixture first. The norm
d
p
(
k
q
)isthespaceofd×k matrices, where we compute
the norm by first taking
q
norm of each row, and then take the
p
norm of the resulting d-
dimensional vector. The strength of such mixed norms has been shown in [14], who showed
that the aforementioned Ulam metric over length-d strings (raised to power 0.9) embeds into
2
[
∞
(
1
)], with dimensions bounded polynomially in d. The distortion of the embedding
is constant. By contrast, embedding into any of the component spaces provably requires
super-constant distortion (see, e.g., Table 7.1 and [67, Chapter 1, Section 8]).
7.5 Sketching
A generalization of the embedding is the notion of sketching. Sketching can be thought
of as a very weak embedding into a computational space, where the host distance is an
arbitrary computation (e.g., not restricted to be a metric). The main parameter is now the
size of the sketch, which can be thought of as host dimension. Sketches address the most
basic, decision version of the distance estimation problem, termed the distance threshold
estimation problem (DTEP) [91]. The goal of DTEP is to distinguish close pairs versus far
pairs with some probability (similar to the Definition 7.3 of near neighbor problem).
Definition 7.5 Fix some metric (M,d
M
), approximation D ≥ 1, threshold r>0,failure
probability δ ∈ [0, 1/3], and size s ∈ N.Themapφ : M →{0, 1}
s
is called a sketch of M if
there exists a referee algorithm R : {0, 1}
s
×{0, 1}
s
→{0, 1}, such that for any x, y ∈ M,
the algorithm R can distinguish between d
M
(x, y) ≤ r (close pair) versus d
M
(x, y) >Dr
(far pair), with probability 1 − δ.
Sketching is useful in many applications, including the NNS problem. In particular, we can
use a sketch of size s (for, say, δ =0.1) to construct an NNS data structure with space
n
O(s)
. The query time is O(log n) times the time to evaluate φ on the query point. Indeed,
from the basic sketch, we can also construct an amplified sketch, which has O(s log n) size
and failure probability at most 1 −1/n
2
: keep k = O(log n) independent copies of the basic
sketch, with the referee algorithm taking the majority vote of those for the k sketches. For
a given query point q, this amplified sketch, termed φ
k
(q), is sufficient to determine the
approximate near neighbor in an n-point dataset P with probability at least 1 − 1/n:run
the referee algorithm on φ
k
(q), φ
k
(p) for each p ∈ P . Because this procedure uses only the
sketch φ
k
(q), of size ks, we can construct an index for each possible ks-bit input, storing
the solution. Overall, we obtain a data structure with 2
ks
= n
O(s)
space.
Sketches have many applications beyond NNS, including in data-streaming algorithms
and compressed sensing, in part by providing an alternative notion of dimension reduction
(as we will see later). More recently, it has found uses to speed up algorithms, for example,
in numerical linear algebra [95].