TY - GEN
T1 - Distributed big-data optimization via block communications
AU - Notarnicola, Ivano
AU - Sun, Ying
AU - Scutari, Gesualdo
AU - Notarstefano, Giuseppe
N1 - Publisher Copyright:
© 2017 IEEE.
PY - 2017/7/2
Y1 - 2017/7/2
N2 - We study distributed multi-agent large-scale optimization problems, wherein the cost function is composed of a smooth possibly nonconvex sum-utility plus a DC (Difference-of-Convex) regularizer. We consider the scenario where the dimension of the optimization variables is so large that optimizing and/or transmitting the entire set of variables could cause unaffordable computation and communication overhead. To address this issue, we propose the first distributed algorithm whereby agents optimize and communicate only a portion of their local variables. The scheme hinges on successive convex approximation (SCA) to handle the nonconvexity of the objective function, coupled with a novel block-signal tracking scheme, aiming at locally estimating the average of the agents' gradients. Asymptotic convergence to stationary solutions of the nonconvex problem is established. Numerical results on a sparse regression problem show the effectiveness of the proposed algorithm and the impact of the block size on its practical convergence speed and communication cost.
AB - We study distributed multi-agent large-scale optimization problems, wherein the cost function is composed of a smooth possibly nonconvex sum-utility plus a DC (Difference-of-Convex) regularizer. We consider the scenario where the dimension of the optimization variables is so large that optimizing and/or transmitting the entire set of variables could cause unaffordable computation and communication overhead. To address this issue, we propose the first distributed algorithm whereby agents optimize and communicate only a portion of their local variables. The scheme hinges on successive convex approximation (SCA) to handle the nonconvexity of the objective function, coupled with a novel block-signal tracking scheme, aiming at locally estimating the average of the agents' gradients. Asymptotic convergence to stationary solutions of the nonconvex problem is established. Numerical results on a sparse regression problem show the effectiveness of the proposed algorithm and the impact of the block size on its practical convergence speed and communication cost.
UR - http://www.scopus.com/inward/record.url?scp=85050757580&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85050757580&partnerID=8YFLogxK
U2 - 10.1109/CAMSAP.2017.8313176
DO - 10.1109/CAMSAP.2017.8313176
M3 - Conference contribution
AN - SCOPUS:85050757580
T3 - 2017 IEEE 7th International Workshop on Computational Advances in Multi-Sensor Adaptive Processing, CAMSAP 2017
SP - 1
EP - 5
BT - 2017 IEEE 7th International Workshop on Computational Advances in Multi-Sensor Adaptive Processing, CAMSAP 2017
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 7th IEEE International Workshop on Computational Advances in Multi-Sensor Adaptive Processing, CAMSAP 2017
Y2 - 10 December 2017 through 13 December 2017
ER -