Communication-Efficient Optimization and Learning for Distributed Multi-agent Systems

dc.contributor.advisorTian, Zhi ZT
dc.creatorXu, Ping
dc.date.accessioned2023-03-17T19:05:27Z
dc.date.available2023-03-17T19:05:27Z
dc.date.issued2022
dc.description.abstractDistributed learning has attracted extensive interest in recent years, owing to the explosion of data generated from mobile sensors, social media services, and other networked multi-agent applications. In many of these applications, the observed data are usually kept private at local sites without being aggregated to a fusion center, either due to the prohibitively high cost of raw data transmission or privacy concerns. Meanwhile, each agent in the network only communicates with its neighbors within a one-hop local range to save transmission power. Moreover, distributed learning is typically implemented in an iterative manner for computational feasibility and efficiency. This incurs frequent communications among agents to exchange their locally computed updates of the shared learning model, which can cause tremendous communication overhead in terms of both link bandwidth and transmission power. Under this circumstance, this dissertation focuses on developing communication-efficient distributed learning algorithms for multi-agent systems under communication and privacy constraints. To be specific, we utilize the random feature (RF) mapping method to circumvent the curse of dimensionality issue in traditional kernel methods and bypass transmitting raw data in distributed kernel learning. This approach enables the reformulation of decentralized kernel learning as a decentralized consensus optimization problem in the RF space, whichis then solved distributedly and iteratively via the alternating direction method of multipliers (ADMM), with a communication-censoring strategy incorporated to evaluate if an update is informative enough to be transmitted. For online streaming data with possibly unknown dynamics, this dissertation develops corresponding adaptive and dynamic decentralized learning approaches to learn the optimal function “on the fly” via linearized ADMM and conventional decentralized ADMM, respectively. Communication censoring and quantization strategies are utilized for both approaches to save communication resources. Finally, the present dissertation offers a unified framework of learning nonlinear input-output maps by bridging the gap in kernel methods and neural networks, which leads to the development of an RF-based deep kernel learning network with multiple learning paths.
dc.format.extent150 pages
dc.format.mediumdoctoral dissertations
dc.identifier.urihttps://hdl.handle.net/1920/13123
dc.language.isoen
dc.rightsCopyright 2022 Ping Xu
dc.rights.urihttps://rightsstatements.org/vocab/InC/1.0
dc.subjectADMM
dc.subjectCommunication-censoring
dc.subjectDistributed multi-agent systems
dc.subjectKernel learning
dc.subjectMachine learning and optimization
dc.subjectRandom feature mapping
dc.subject.keywordsEngineering
dc.titleCommunication-Efficient Optimization and Learning for Distributed Multi-agent Systems
dc.typeText
thesis.degree.disciplineElectrical and Computer Engineering
thesis.degree.grantorGeorge Mason University
thesis.degree.levelDoctoral
thesis.degree.namePh.D. in Electrical and Computer Engineering

Files

Original bundle
Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
Xu_gmu_0883E_12820.pdf
Size:
4.52 MB
Format:
Adobe Portable Document Format