谷歌浏览器插件
订阅小程序
在清言上使用

Near neighbor preserving dimension reduction for doubling subsets of $\ell_1$.

International Workshop and International Workshop on Approximation, Randomization, and Combinatorial Optimization. Algorithms and Techniques(2019)

引用 23|浏览7
暂无评分
摘要
Randomized dimensionality reduction has been recognized as one of the fundamental techniques in handling high-dimensional data. Starting with the celebrated Johnson-Lindenstrauss Lemma, such reductions have been studied in depth for the Euclidean $(ell_2)$ metric and, much less, for the Manhattan $(ell_1)$ metric. Our primary motivation is the approximate nearest neighbor problem in $ell_1$. We exploit its reduction to the decision-with-witness version, called approximate textit{near} neighbor, which incurs a roughly logarithmic overhead. In 2007, Indyk and Naor, in the context of approximate nearest neighbors, introduced the notion of nearest neighbor-preserving embeddings. These are randomized embeddings between two metric spaces with guaranteed bounded distortion only for the distances between a query point and a point set. Such embeddings are known to exist for both $ell_2$ and $ell_1$ metrics, as well as for doubling subsets of $ell_2$. In this paper, we propose a dimension reduction, textit{near} neighbor-preserving embedding for doubling subsets of $ell_1$. Our approach is to represent the point set with a carefully chosen covering set, and then apply a random projection to that covering set. We study two cases of covering sets: $c$-approximate $r$-nets and randomly shifted grids, and we discuss the tradeoff between them in terms of preprocessing time and target dimension.
更多
查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要