Parameterized Approximation Schemes for Clustering with General Norm Objectives

2023 IEEE 64TH ANNUAL SYMPOSIUM ON FOUNDATIONS OF COMPUTER SCIENCE, FOCS(2023)

引用 2|浏览25
暂无评分
摘要
This paper considers the well-studied algorithmic egime of designing a (1+epsilon)-approximation algorithm for a k-clustering problem that runs in time f(k, epsilon)poly(n) (sometimes called an efficient parameterized approximation scheme or EPAS for short(1)). Notable results of this kind include EPASes in the high-dimensional Euclidean setting for k-center [Badoiu, Har-Peled, Indyk; STOC'02] as well as k-median, and k-means [Kumar, Sabharwal, Sen; J. ACM 2010]. Our main contribution is a clean and simple EPAS that settles more than ten clustering problems (across multiple well-studied objectives as well as metric spaces) and unifies well-known EPASes. More specifically, our algorithm gives EPASes in the following settings: Clustering objectives: k-means, k-center, k-median, priority k-center, l-centrum, ordered k-median, socially fair k-median (aka robust k-median), or any other objective that can be formulated as minimizing a monotone (not necessarily symmetric!) norm of the distances of the points from the solution (generalizing the symmetric formulation introduced by Chakrabarty and Swamy [STOC'19]). Metric spaces: Continuous high-dimensional Euclidean spaces, metrics of bounded doubling dimension, bounded treewidth metrics, and planar metrics. Prior to our results, EPASes were only known for vanilla clustering objectives (k-means, k-median, and k-center) and each such algorithm is tailored to work for the specific input metric and clustering objective (e.g., EPASes for k-means and k-center in R-d are conceptually very different). In contrast, our algorithmic framework is applicable to a wide range of well-studied objective functions in a uniform way, and is (almost) entirely oblivious to any specific metric structures and yet is able to effectively exploit those unknown structures. In particular, our algorithm is not based on the (metric- and objective-specific) technique of coresets. Key to our analysis is a new concept that we call bounded epsilon-scatter dimension-an intrinsic complexity measure of a metric space that is a relaxation of the standard notion of bounded doubling dimension (often used as a source of algorithmic tractability for geometric problems). Our main technical result shows that two conditions are essentially sufficient for our algorithm to yield an EPAS on the input metric M for any clustering objective: (i) The objective is described by a monotone norm, and (ii) the epsilon-scatter dimension of M is upper bounded by a function of epsilon.
更多
查看译文
关键词
clustering,parameterized approximation algorithms,scattered dimension,norm clustering
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要