Faster, Incentivized, and Efficient Federated Learning: Theory and Applications
Artificial intelligence (AI) is becoming increasingly ubiquitous with a plethora of applications such as recommendation systems, image/video generation, or chatbots changing our modern society. The majority of these AI based applications are based on centralized machine learning (ML). In centralized ML, a large amount of data is collected at a central server to train a large ML model with hundreds of billions of parameters using massive amount of system resources (e.g., CPUs, GPUs). Despite the success of such pipeline for training models in the centralized setting, however, centralized ML has critical limitations. First, data collection can raise serious privacy concerns due to the desired data often being personal data such as medical histories or financial data and users opting out from sharing their data. Second, the exponentially increasing economic and environmental cost for training colossal models in the centralized setting raises concerns for sustainability.
Such major drawbacks of centralized ML calls for a new paradigm that can shift the server-based ML pipeline to the edge, where data collection as well as computation happens on-device at the edge client. Moreover, decentralizing ML to the edge can facilitate democratizing the benefits of ML by allowing better personalized models for individual users. Federated learning (FL) is a well-known method that achieves such decentralization of ML where clients (e.g., mobile phones, hospitals, banks) locally train the server ML model(s) on their private data and send only the local gradient updates to the server so that the server can update its model with the aggregation of these local updates. The clients can also privately further fine-tune the server ML model to make personalized models to their data.
In this thesis, we aim to address the three main challenges that arise in FL which does not exist in the conventional centralized ML setting: i) clients’ limited communication and availability, ii) clients’ data heterogeneity, and iii) clients’ limited and heterogeneous training resources. We address each challenge with three main research contributions respectively i) faster FL, ii) incentivized FL, and iii) efficient FL. First, we investigate theoretically and empirically how biased client selection and cyclic client participation in FL under limited client communication and availability can improve the convergence of FL for faster FL. Second, we delve into the question of whether clients have incentives to join FL in the first place under client data heterogeneity and how disincentivization of clients can sabotage the training process of FL. We propose a framework to improve the fraction of incentivized clients by incentive-aware weighted aggregation of clients’ local updates for incentivized FL and also propose a communication-efficient personalized FL framework to further incentivize clients. Lastly, we propose two novel methods, federated ensemble transfer and heterogeneous LoRA (low-rank approximation) to train larger server models, including foundation models, at the edge with significantly smaller number of parameters under clients’ limited and heterogeneous training resources. We also look into efficient FL from the perspective of labels where clients have only a few labels in their local data.
History
Date
2024-05-22Degree Type
- Dissertation
Department
- Electrical and Computer Engineering
Degree Name
- Doctor of Philosophy (PhD)