This paper considers the problem of distributed multi-agent learning, where the global aim is to minimize a sum of local objective (empirical loss) functions through local optimization and information exchange between neighboring nodes. We introduce a Newton-type fully distributed optimization algorithm, Network-GIANT, which is based on GIANT, a Federated learning algorithm that relies on a central parameter server. Network-GIANT is designed via a combination of gradient-tracking and a Newton-type iterative algorithm at each node, with consensus-based averaging of Newton updates. The resulting algorithm is efficient in terms of both communication cost and runtime, making it suitable for wireless networks. We prove that our algorithm guarantees semi-global and exponential convergence to the exact solution over the network assuming strongly convex and smooth loss functions. We provide empirical evidence of the superior convergence performance of Network-GIANT over other state-of-art distributed learning algorithms such as Network-DANE and Newton-Raphson Consensus.
Network-GIANT: Fully Distributed Newton-Type Optimization via Harmonic Hessian Consensus
Maritan A.;Schenato L.;
2023
Abstract
This paper considers the problem of distributed multi-agent learning, where the global aim is to minimize a sum of local objective (empirical loss) functions through local optimization and information exchange between neighboring nodes. We introduce a Newton-type fully distributed optimization algorithm, Network-GIANT, which is based on GIANT, a Federated learning algorithm that relies on a central parameter server. Network-GIANT is designed via a combination of gradient-tracking and a Newton-type iterative algorithm at each node, with consensus-based averaging of Newton updates. The resulting algorithm is efficient in terms of both communication cost and runtime, making it suitable for wireless networks. We prove that our algorithm guarantees semi-global and exponential convergence to the exact solution over the network assuming strongly convex and smooth loss functions. We provide empirical evidence of the superior convergence performance of Network-GIANT over other state-of-art distributed learning algorithms such as Network-DANE and Newton-Raphson Consensus.Pubblicazioni consigliate
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.




