TY - GEN
T1 - Stochastic gradient descent with only one projection
AU - Mahdavi, Mehrdad
AU - Yang, Tianbao
AU - Jin, Rong
AU - Zhu, Shenghuo
AU - Yi, Jinfeng
PY - 2012
Y1 - 2012
N2 - Although many variants of stochastic gradient descent have been proposed for large-scale convex optimization, most of them require projecting the solution at each iteration to ensure that the obtained solution stays within the feasible domain. For complex domains (e.g., positive semidefinite cone), the projection step can be computationally expensive, making stochastic gradient descent unattractive for large-scale optimization problems. We address this limitation by developing novel stochastic optimization algorithms that do not need intermediate projections. Instead, only one projection at the last iteration is needed to obtain a feasible solution in the given domain. Our theoretical analysis shows that with a high probability, the proposed algorithms achieve an O(1/√T) convergence rate for general convex optimization, and an O (ln T/T) rate for strongly convex optimization under mild conditions about the domain and the objective function.
AB - Although many variants of stochastic gradient descent have been proposed for large-scale convex optimization, most of them require projecting the solution at each iteration to ensure that the obtained solution stays within the feasible domain. For complex domains (e.g., positive semidefinite cone), the projection step can be computationally expensive, making stochastic gradient descent unattractive for large-scale optimization problems. We address this limitation by developing novel stochastic optimization algorithms that do not need intermediate projections. Instead, only one projection at the last iteration is needed to obtain a feasible solution in the given domain. Our theoretical analysis shows that with a high probability, the proposed algorithms achieve an O(1/√T) convergence rate for general convex optimization, and an O (ln T/T) rate for strongly convex optimization under mild conditions about the domain and the objective function.
UR - http://www.scopus.com/inward/record.url?scp=84877725845&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=84877725845&partnerID=8YFLogxK
M3 - Conference contribution
AN - SCOPUS:84877725845
SN - 9781627480031
T3 - Advances in Neural Information Processing Systems
SP - 494
EP - 502
BT - Advances in Neural Information Processing Systems 25
T2 - 26th Annual Conference on Neural Information Processing Systems 2012, NIPS 2012
Y2 - 3 December 2012 through 6 December 2012
ER -