Exact Diffusion for Distributed Optimization and Learning-Part II: Convergence Analysis

Part I of this paper developed the exact diffusion algorithm to remove the bias that is characteristic of distributed solutions for deterministic optimization problems. The algorithm was shown to be applicable to the larger set of locally balanced left-stochastic combination policies than the set of...

Full description

Saved in:
Bibliographic Details
Published in:IEEE transactions on signal processing Vol. 67; no. 3; pp. 724 - 739
Main Authors: Yuan, Kun, Ying, Bicheng, Zhao, Xiaochuan, Sayed, Ali H.
Format: Journal Article
Language:English
Published: New York IEEE 01-02-2019
The Institute of Electrical and Electronics Engineers, Inc. (IEEE)
Subjects:
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:Part I of this paper developed the exact diffusion algorithm to remove the bias that is characteristic of distributed solutions for deterministic optimization problems. The algorithm was shown to be applicable to the larger set of locally balanced left-stochastic combination policies than the set of doubly-stochastic policies. These balanced policies endow the algorithm with faster convergence rate, more flexible step-size choices and better privacy-preserving properties. In this Part II, we examine the convergence and stability properties of exact diffusion in some detail and establish its linear convergence rate. We also show that it has a wider stability range than the EXTRA consensus solution, meaning that it is stable for a wider range of step-sizes and can, therefore, attain faster convergence rates. Analytical examples and numerical simulations illustrate the theoretical findings.
ISSN:1053-587X
1941-0476
DOI:10.1109/TSP.2018.2875883