log in  |  register  |  feedback?  |  help  |  web accessibility
From Learning through Labels to Learning through Language
Kai-Wei Chang
4105
Tuesday, February 25, 2025, 2:30-3:30 pm
  • You are subscribed to this talk through .
  • You are watching this talk through .
  • You are subscribed to this talk. (unsubscribe, watch)
  • You are watching this talk. (unwatch, subscribe)
  • You are not subscribed to this talk. (watch, subscribe)
Abstract

Over the past decades, machine learning has primarily relied on labeled data, with success often depending on the availability of vast, high-quality annotations and the assumption that test conditions mirror training conditions. In contrast, humans learn efficiently from conceptual explanations, instructions, rules, and contextual understanding. With advancements in large language models, AI systems can now understand descriptions and follow instructions, paving the way for a paradigm shift. Inspired by this, we investigate how to empower AI agents to learn from natural language narratives and multimodal interactions with humans, making them more adaptable to rapidly changing environments. 

This talk explores how teaching machines through language can enable AI systems to gain human trust and enhance their interpretability, robustness, and ability to learn new concepts. I will highlight our journey in developing vision-language models capable of detecting unseen objects through rich natural language descriptions. Additionally, I will discuss techniques for guiding the behavior of language models and text-to-image models using language, and how we generate human-readable features using vision-language models. Finally, I will conclude the talk by discussing future directions and potential challenges in empowering models to learn through language.

Bio

Kai-Wei Chang is an Associate Professor in the Department of Computer Science at UCLA and an Amazon Scholar at Amazon AGI. His research interests include designing trustworthy natural language processing systems and developing multimodal models for vision-language applications. Kai-Wei has published broadly in NLP, AI, and ML. His awards include IEEE AI's 10 to Watch (2024), the Sloan Fellow (2021), AAAI Senior Member (2023), CVPR Best Paper Finalist (2022), EMNLP Best Long Paper Award (2017), and KDD Best Paper Award (2010).

Kai-Wei was elected as an officer of SIGDAT, the organizing body behind EMNLP, and will serve as President in 2026. He is an associate editor for journals such as JAIR, JMLR, TACL, and ARR and senior area chair for most ML and NLP conferences. He has delivered multiple tutorials on topics such as Fairness, Robustness, and Multimodal NLP at EMNLP (2019, 2021) and ACL (2023). Kai-Wei received his Ph.D. from the University of Illinois at Urbana-Champaign in 2015 and subsequently worked as a postdoctoral researcher at Microsoft Research in 2016. For more details, visit http://kwchang.net.

 

This talk is organized by Samuel Malede Zewdu