Part of Advances in Neural Information Processing Systems 36 (NeurIPS 2023) Main Conference Track
Hongjie Chen, Vincent Cohen-Addad, Tommaso d’Orsi, Alessandro Epasto, Jacob Imola, David Steurer, Stefan Tiegel
We introduce general tools for designing efficient private estimation algorithms, in the high-dimensional settings, whose statistical guarantees almost match those of the best known non-private algorithms.To illustrate our techniques, we consider two problems: recovery of stochastic block models and learning mixtures of spherical Gaussians.For the former, we present the first efficient (ϵ,δ)-differentially private algorithm for both weak recovery and exact recovery. Previously known algorithms achieving comparable guarantees required quasi-polynomial time. For the latter, we design an (ϵ,δ)-differentially private algorithm that recovers the centers of the k-mixture when the minimum separation is at least O(k1/t√t). For all choices of t, this algorithm requires sample complexity n≥kO(1)dO(t) and time complexity (nd)O(t). Prior work required either an additional additive Ω(√logn) term in the minimum separation or an explicit upper bound on the Euclidean norm of the centers.