log in  |  register  |  feedback?  |  help  |  web accessibility
Logo
Modern spectral theory meets deep learning: nonconvex learning with limited resources
Furong Huang
Virtual-https://umd.zoom.us/j/93637673064?pwd=TzJYcE15UXg0MTJSQXJ5UFFLMlBNZz09
Friday, October 23, 2020, 11:00 am-12:00 pm Calendar
  • You are subscribed to this talk through .
  • You are watching this talk through .
  • You are subscribed to this talk. (unsubscribe, watch)
  • You are watching this talk. (unwatch, subscribe)
  • You are not subscribed to this talk. (watch, subscribe)
Abstract

Optimizing non-convex objective functions is central to machine learning (ML), such as modern deep learning and classical latent variable model learning. Spectral methods that go beyond matrices tackle the problem of non-convexity in learning latent variable models and provide provable performance guarantees using tensor decompositions. In contrast, the popular gradient-based optimization methods for deep neural networks often lack convergence guarantees. Our objective is to develop modern spectral theory - generalized tensor decompositions - that advances existing spectral methods to be adaptable for learning deep neural networks under limited resources. The talk focuses on non-convex learning with limited resources, a central challenge in modern ML settings as federated learning and personalized ML on constrained devices gain increasing popularity.

 

An overview of our group’s research progress on performance guaranteed spectral methods for deep neural networks will be presented. Using tensor representation theory, we extend traditional neural nets to more general tensorial neural networks that are interpretable and compact. We demonstrate tensorial neural networks' advantages over traditional neural networks by theoretically proving their improved generalization and expressive power with fewer parameters. Using the developed new class of latent variable model learning, we establish guaranteed learning methods for deep ResNets. Finally, we obtain certifiably robust deep networks using latent variables and architecture design in the spectral domain.

 

Bio

Dr. Furong Huang is an Assistant Professor of the Department of Computer Science at the University of Maryland. She is a recipient of the MLconf Industry Impact Research Award, the Adobe Faculty Research Award, two JP Morgan Faculty Research Awards and NSF CRII Award. Her research focuses on machine learning, high-dimensional statistics, nonconvex optimization, spectral methods and deep learning theory. Dr Huang’s recent work on deep learning theory involves guaranteed training of neural network agnostic to data distribution which includes adversarial examples, understanding generalization ability of neural network via compressibility of neural nets using tensor algebra and implementing practical approaches guided by generalization theory for label corrupted data. 

This talk is organized by Richa Mathur