PhD Defense: Robust Learning under Distributional Shifts

Talk
Yogesh Balaji
Time: 
06.10.2021 14:00 to 16:00
Location: 

Remote

Designing robust models is critical for the reliable deployment of artificial intelligence systems. Deep neural networks perform exceptionally well on test samples that are drawn from the same distribution as the training set. However, they perform poorly when there is a mismatch between training and test conditions, a phenomenon called distributional shift. For instance, the perception system of a self-driving car can produce erratic predictions when it encounters a test sample with a different illumination or weather condition not seen during training. Such inconsistencies are undesirable, and can potentially create life-threatening conditions when these models are deployed in safety-critical applications.In this dissertation, we develop several techniques for effectively handling distributional shifts in deep learning systems.In the first part of the dissertation, we focus on detecting out-of-distribution shifts that can be used for flagging outlier samples at test-time. We develop a likelihood estimation framework based on deep generative models for this task. In the second part, we study the domain adaptation problem where the objective is to tune the neural network models to adapt to a specific target distribution of interest. We design novel adaptation algorithms based on variants of optimal transport distances, understand and analyze them under various settings. Finally, we focus on designing robust learning algorithms that can generalize to novel test-time distributional shifts. In particular, we study two types of distributional shifts: covariate and adversarial shifts. All developed algorithms are rigorously evaluated on several benchmark computer vision datasets.
Examining Committee:

Chair: Dr. Rama Chellappa Co-Chair: Dr. Soheil Feizi Dean's rep: Dr. Wojciech Czaja Members: Dr. Abhinav Shrivastava
Dr. Tom Goldstein