Journal
SIAM JOURNAL ON OPTIMIZATION
Volume 27, Issue 2, Pages 1171-1203Publisher
SIAM PUBLICATIONS
DOI: 10.1137/15M1038049
Keywords
distributed optimization; second order methods; Newton-like methods; linear convergence
Categories
Funding
- Serbian Ministry of Education, Science, and Technological Development [174030]
Ask authors/readers for more resources
We consider distributed optimization problems where networked nodes cooperatively minimize the sum of their locally known convex costs. A popular class of methods to solve these problems are the distributed gradient methods, which are attractive due to their inexpensive iterations, but have a drawback of slow convergence rates. This motivates the incorporation of second order information in the distributed methods, but this task is challenging: although the Hessians which arise in the algorithm design respect the sparsity of the network, their inverses are dense, hence rendering distributed implementations difficult. We overcome this challenge and propose a class of distributed Newton-like methods, which we refer to as distributed quasi-newton (DQN). The DQN family approximates the Hessian inverse by (1) splitting the Hessian into its diagonal and off diagonal parts, (2) inverting the diagonal part, and (3) approximating the inverse of the off-diagonal part through a weighted linear function. The approximation is parameterized by the tuning variables which correspond to different splittings of the Hessian and by different weightings of the off-diagonal Hessian part. Specific choices of the tuning variables give rise to different variants of the proposed general DQN method dubbed DQN-0, DQN-1, and DQN-2 which mutually trade-off communication and computational costs for convergence. Simulations demonstrate the effectiveness of the proposed DQN methods.
Authors
I am an author on this paper
Click your name to claim this paper and add it to your profile.
Reviews
Recommended
No Data Available