Notice and Invitation
Oral Defense of Doctoral Dissertation
The College of Engineering and Computing, George Mason University
Ping Xu
Bachelor of Science, Northwestern Polytechnical University, Xi’an, China, 2015
Master of Science, George Mason University, Fairfax, VA, 2018
Communication-efficient optimization and learning for distributed multi-agent systems
Monday, April 18, 2022
10:30 AM – 12:00 PM
ENGR 2901
All are invited to attend.
Committee
Dr. Zhi (Gerry) Tian, Chair
Dr. Yariv Ephraim
Dr. Kathleen E. Wage
Dr. Kuo-Chu Chang
Abstract
Distributed learning has attracted extensive interest in recent years, owing to the explosion of data generated from mobile sensors, social media services, and other networked multi-agent applications. In many of these applications, the observed data are usually kept private at local sites without being aggregated to a fusion center, either due to the prohibitively high cost of raw data transmission or privacy concerns. Meanwhile, each agent in the network only communicates with its one-hop neighbors within its local area to save transmission power. Moreover, distributed learning is typically implemented in an iterative manner for computational feasibility and efficiency. This incurs frequent communications among agents to exchange their locally computed updates of the shared learning model, which can cause tremendous communication overhead in terms of both link bandwidth and transmission power. Under this circumstance, this dissertation focuses on developing communication-efficient distributed learning algorithms for multi-agent systems under communication and privacy constraints.
To be specific, we utilize the random feature (RF) mapping method to circumvent the curse of dimensionality issue in traditional kernel methods and bypass transmitting raw data in distributed kernel learning. This approach enables the reformulation of decentralized kernel learning as a decentralized consensus optimization problem in the RF space, which is then solved distributedly and iteratively via the alternating direction method of multipliers (ADMM), with a communication-censoring strategy incorporated to evaluate if an update is informative enough to be transmitted. For online streaming data with possibly unknown dynamics, this dissertation develops corresponding adaptive and dynamic decentralized learning approaches to learn the optimal function ``on the fly" via linearized ADMM, and conventional decentralized ADMM, respectively. Communication censoring and quantization strategies are utilized for both approaches to save communication resources. Finally, the present dissertation offers a unified framework of learning from nonlinear input-output maps by bridging the gap in kernel methods and neural networks, as well as the development of a RF-based deep kernel learning network with multiple learning path.
|