Emiris, Ioannis Z. ;
Margonis, Vasilis ;
Psarros, Ioannis
NearNeighbor Preserving Dimension Reduction for Doubling Subsets of l_1
Abstract
Randomized dimensionality reduction has been recognized as one of the fundamental techniques in handling highdimensional data. Starting with the celebrated JohnsonLindenstrauss Lemma, such reductions have been studied in depth for the Euclidean (l_2) metric, but much less for the Manhattan (l_1) metric. Our primary motivation is the approximate nearest neighbor problem in l_1. We exploit its reduction to the decisionwithwitness version, called approximate near neighbor, which incurs a roughly logarithmic overhead. In 2007, Indyk and Naor, in the context of approximate nearest neighbors, introduced the notion of nearest neighborpreserving embeddings. These are randomized embeddings between two metric spaces with guaranteed bounded distortion only for the distances between a query point and a point set. Such embeddings are known to exist for both l_2 and l_1 metrics, as well as for doubling subsets of l_2. The case that remained open were doubling subsets of l_1. In this paper, we propose a dimension reduction by means of a near neighborpreserving embedding for doubling subsets of l_1. Our approach is to represent the pointset with a carefully chosen covering set, then randomly project the latter. We study two types of covering sets: capproximate rnets and randomly shifted grids, and we discuss the tradeoff between them in terms of preprocessing time and target dimension. We employ Cauchy variables: certain concentration bounds derived should be of independent interest.
BibTeX  Entry
@InProceedings{emiris_et_al:LIPIcs:2019:11262,
author = {Ioannis Z. Emiris and Vasilis Margonis and Ioannis Psarros},
title = {{NearNeighbor Preserving Dimension Reduction for Doubling Subsets of l_1}},
booktitle = {Approximation, Randomization, and Combinatorial Optimization. Algorithms and Techniques (APPROX/RANDOM 2019)},
pages = {47:147:13},
series = {Leibniz International Proceedings in Informatics (LIPIcs)},
ISBN = {9783959771252},
ISSN = {18688969},
year = {2019},
volume = {145},
editor = {Dimitris Achlioptas and L{\'a}szl{\'o} A. V{\'e}gh},
publisher = {Schloss DagstuhlLeibnizZentrum fuer Informatik},
address = {Dagstuhl, Germany},
URL = {http://drops.dagstuhl.de/opus/volltexte/2019/11262},
URN = {urn:nbn:de:0030drops112628},
doi = {10.4230/LIPIcs.APPROXRANDOM.2019.47},
annote = {Keywords: Approximate nearest neighbor, Manhattan metric, randomized embedding}
}
2019
Keywords: 

Approximate nearest neighbor, Manhattan metric, randomized embedding 
Seminar: 

Approximation, Randomization, and Combinatorial Optimization. Algorithms and Techniques (APPROX/RANDOM 2019)

Issue date: 

2019 
Date of publication: 

2019 