log in  |  register  |  feedback?  |  help  |  web accessibility
Logo
PhD Defense: Foundations of Trustworthy Deep Learning: Fairness, Robustness, and Explainability
Vedant Nanda
Monday, April 1, 2024, 2:00-3:00 pm Calendar
  • You are subscribed to this talk through .
  • You are watching this talk through .
  • You are subscribed to this talk. (unsubscribe, watch)
  • You are watching this talk. (unwatch, subscribe)
  • You are not subscribed to this talk. (watch, subscribe)
Abstract

Deep Learning (DL) models, especially with the rise of the so-called foundation models, are increasingly used in real-world applications either as autonomous systems (eg: facial recognition), as decision aids (eg: medical imaging, writing assistants), and even to generate novel content (eg: chatbots, image generators). This naturally results in concerns about the trustworthiness of these systems, for example, do the models systematically perform worse for certain subgroups? Are the outputs of these models reliable under perturbations to the inputs? This thesis aims to strengthen the foundations of DL models, so they can be trusted in deployment. I will cover three important aspects of trust: fairness, robustness, and explainability. I will argue that we need to expand the scope of each of these aspects when applying them to DL models and carefully consider possible tradeoffs between these desirable but sometimes conflicting notions of trust.

In the first part, I will present two works that show how thinking about fairness for deep learning (DL) introduces new challenges, especially due to their overparametrized nature and susceptibility to adversarial attacks. In the second part, I will argue that to get truly robust models, we must focus on a more general notion of robustness: measuring the alignment of invariances of DL models with other models of perception such as humans. Finally, in the last part, I will show how even a small subset of randomly chosen neurons from a pre-trained representation can transfer very well to downstream tasks which challenges existing beliefs in the explainability literature that claim individual neurons learn disjoint semantically meaningful concepts.

Bio

Vedant Nanda is a final year PhD student and is part of the Maryland Max Planck joint program. He is advised by Profs. John P. Dickerson (University of Maryland) and Krishna P. Gummadi (MPI-SWS). His research focuses on better understanding deep learning through the lens of fairness, robustness, explainability, and efficiency.

This talk is organized by Migo Gui