log in  |  register  |  feedback?  |  help  |  web accessibility
Logo
Toward Practical and Scalable ML Safety
Friday, November 3, 2023, 11:00 am-12:00 pm Calendar
  • You are subscribed to this talk through .
  • You are watching this talk through .
  • You are subscribed to this talk. (unsubscribe, watch)
  • You are watching this talk. (unwatch, subscribe)
  • You are not subscribed to this talk. (watch, subscribe)
Abstract

zoom: https://umd.zoom.us/j/99147773612

 

Despite recent breakthroughs in aid of deep learning and scaling laws, neural networks still make a substantially fragile prediction in their worst-case behaviors, e.g., as shown in their performance degradation to adversarial examples, novelties, and corruptions. Literatures have been independently developed to address such different types of distribution shifts, they often result in bespoke models to their threat models during training, eventually limiting their applicability to recent models at scale. In this talk, I will present some of my attempts in order to make ML safety more practical and scalable, covering multiple aspects of ML safety across robustness, monitoring and alignment research. Firstly, I will introduce an Information Bottleneck based approach to mitigate the train-time dependency of robustness-aware training methods, i.e., on specific data augmentation techniques. Next, we move on our focus to adversarial robustness, and consider a crucial problem of robustness-accuracy trade-off: with an introduction of a recent technique of randomized smoothing, I will introduce a simple and scalable idea to overcome the trade-off. Lastly, I will briefly cover some of my recent works that highlights the effectiveness of language-based inference to perform efficient monitoring and uncertainty estimation.

Bio

Jongheon is a postdoctoral researcher at Korea Advanced Institute of Science and Technology (KAIST). He obtained his Ph.D. at KAIST in August 2023, where he was advised by Prof. Jinwoo Shin. During his Ph.D. studies, he interned at Amazon Web Services (AWS)(Seattle, WA) twice, in 2022 and 2021. He is also a recipient of Qualcomm Innovation Fellowship Korea 2020 from two of his papers. Previously, he received a B.S. in Mathematics and Computer Science from KAIST in 2017.

Jongheon is broadly interested in discovering (if exist) simple priors that would close the gap between neural network and human perception. Many topics are related, particularly on (but not limited to) robustness (or generalization) against distribution shifts, e.g., adversarial examples, natural corruptions, out-of-distribution, and label shifts, to name a few. Ultimately, his research aims to understand why neural networks behave so differently from our brain, and how our brain makes such reliable yet efficient inferences.

This talk is organized by Hal Daume III