Carnegie Mellon University
Browse

Robust Machine Learning: Detection, Evaluation and Adaptation Under Distribution Shift

Download (16.09 MB)
thesis
posted on 2024-07-23, 17:01 authored by Saurabh GargSaurabh Garg

Deep learning, despite its broad applicability, grapples with robustness chal lenges in real-world applications, especially when training and test distributions  differ. Reasons for the discrepancy between training and test distributions  include gradual changes in human behavior or differences in the demographics  of the environment where the service is being used. While obtaining labeled  data for anticipated distribution shifts can be daunting, unlabeled samples are  relatively cheap and abundantly available My research leverages unlabeled data from the target domain to identify  structural relationships between the target and source domains, and then use  them to adapt and evaluate models. The work discussed in thesis involves  understanding the behavior of deep models, both theoretically and empirically,  and using those insights to develop robust methods. In particular, this thesis  surveys my work on the following three questions:  

Q1: How to adapt models in the face of distribution shifts? Absent assump tions on the nature of the distribution shifts, this task is impossible. My research  in this direction is focused on formulating assumptions on distribution shift  scenarios appearing in the wild and developing procedures that improve and  adapt deep models under those shifts by leveraging unlabeled data. Part I and  II of this thesis delve into this research.  

Q2: How can we evaluate models’ performance without access to labeled data?  Deep learning models fail silently, i.e., they cannot flag uncertain decisions. To  build reliable machine learning systems, obtaining certificates for accuracy is as  important as robustifying these systems. Part III discusses my research in this  direction and presents techniques that leverage unlabeled data to predict model  accuracy.  

Q3: How can we leverage foundation models to address distribution shift  challenges? Foundation models, such as vision language models, have demon strated impressive performance on a wide range of tasks. However, these models  also lack robustness due to spurious correlations, poor image-text alignment,  etc. Moreover, they also get outdated as the internet data evolves presenting  novel challenges in keeping them up-to-date. Part IV of my thesis discusses  my work on understanding the behavior of foundation models and developing  techniques to improve their robustness under distribution shifts.  

Overall, this thesis expands the frontier of robust machine learning by  developing techniques that leverage unlabeled data to adapt and evaluate  models under distribution shifts. The work presented here is a step towards  developing a comprehensive toolkit for robust machine learning in the face of  distribution shifts.  

History

Date

2024-05-01

Degree Type

  • Dissertation

Department

  • Machine Learning

Degree Name

  • Doctor of Philosophy (PhD)

Advisor(s)

Zachary Lipton Sivaraman Balakrishnan