TY - JOUR
T1 - Sketching meets random projection in the dual
T2 - A provable recovery algorithm for big and high-dimensional data
AU - Wang, Jialei
AU - Lee, Jason D.
AU - Mahdavi, Mehrdad
AU - Kolar, Mladen
AU - Srebro, Nathan
N1 - Publisher Copyright:
© 2017, Institute of Mathematical Statistics. All rights reserved.
PY - 2017
Y1 - 2017
N2 - Sketching techniques scale up machine learning algorithms by reducing the sample size or dimensionality of massive data sets, without sacrificing their statistical properties. In this paper, we study sketching from an optimization point of view. We first show that the iterative Hessian sketch is an optimization process with preconditioning and develop an accelerated version using this insight together with conjugate gradient descent. Next, we establish a primal-dual connection between the Hessian sketch and dual random projection, which allows us to develop an accelerated iterative dual random projection method by applying the preconditioned conjugate gradient descent on the dual problem. Finally, we tackle the problems of large sample size and high-dimensionality in massive data sets by developing the primal-dual sketch. The primal-dual sketch iteratively sketches the primal and dual formulations and requires only a logarithmic number of calls to solvers of small sub-problems to recover the optimum of the original problem up to arbitrary precision. Our iterative sketching techniques can also be applied for solving distributed optimization problems where data are partitioned by samples or features. Experiments on synthetic and real data sets complement our theoretical results.
AB - Sketching techniques scale up machine learning algorithms by reducing the sample size or dimensionality of massive data sets, without sacrificing their statistical properties. In this paper, we study sketching from an optimization point of view. We first show that the iterative Hessian sketch is an optimization process with preconditioning and develop an accelerated version using this insight together with conjugate gradient descent. Next, we establish a primal-dual connection between the Hessian sketch and dual random projection, which allows us to develop an accelerated iterative dual random projection method by applying the preconditioned conjugate gradient descent on the dual problem. Finally, we tackle the problems of large sample size and high-dimensionality in massive data sets by developing the primal-dual sketch. The primal-dual sketch iteratively sketches the primal and dual formulations and requires only a logarithmic number of calls to solvers of small sub-problems to recover the optimum of the original problem up to arbitrary precision. Our iterative sketching techniques can also be applied for solving distributed optimization problems where data are partitioned by samples or features. Experiments on synthetic and real data sets complement our theoretical results.
UR - http://www.scopus.com/inward/record.url?scp=85038369216&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85038369216&partnerID=8YFLogxK
U2 - 10.1214/17-EJS1334SI
DO - 10.1214/17-EJS1334SI
M3 - Article
AN - SCOPUS:85038369216
SN - 1935-7524
VL - 11
SP - 4896
EP - 4944
JO - Electronic Journal of Statistics
JF - Electronic Journal of Statistics
IS - 2
ER -