log in  |  register  |  feedback?  |  help  |  web accessibility
Logo
PhD Proposal: Move Faster or Go Smaller: Efficient Models and Learning Paradigms for Resource-Constrained Systems
Tahseen Rabbani
Tuesday, April 25, 2023, 3:30-5:30 pm Calendar
  • You are subscribed to this talk through .
  • You are watching this talk through .
  • You are subscribed to this talk. (unsubscribe, watch)
  • You are watching this talk. (unwatch, subscribe)
  • You are not subscribed to this talk. (watch, subscribe)
Abstract
As deep learning algorithms continue to grow more ambitious in task complexity, so do the sizes of their underlying architectures and training complexities. The ability to host and quickly compute powerful models on more modest devices, such as mobile devices, sensors, smartwatches, etc. is a necessary research question to consider as Internet of Things (IoT) technologies continue to expand in technology and demand. Motivated by these challenges, we consider two approaches to improving efficiency: (1) reducing model size and (2) minimizing the amount of computation and time spent training. Emerging frameworks such as federated learning (FL), graph neural networks (GNNs), and tensorial neural networks (TNNs) impose additional privacy and resource constraints which require novel variations upon classical methods and entirely new strategies altogether.

The first approach (1) involves dimensionality reduction of taxing model structures. We first propose a novel Count Sketch parametrization of fully-connected and convolutional weights which is desirable for small clients operating in an FL ecosystem. We then apply Count Sketch compression to GNNs to sketch nonlinear activations and node embeddings, resulting in a message passing protocol which achieves sublinear training time and memory complexities with respect to graph size. Returning to FL and inspired by recent literature in adaptive sparsification/pruning via locality-sensitive hashing (LSH), we develop a new hashing family, the Periodic Gaussian Hash (PGHash), which enables federated clients to dynamically prune massive, fully-connected weights via private, on-device LSH.

The second approach (2) seeks to reduce the computational and time complexity of various learning paradigms through an ad hoc mixture of momentum, automated machine learning (AutoML), and asynchronous methods. We first accelerate the principal component analysis (PCA) underlying unsupervised spectral clustering via our novel variation of the classical power method, DMPower, which incorporates a semi-adaptive momentum term. Shifting our focus to TNNs, we develop an automated framework and open-source library, AutoTNN, for compressing and minimizing floating point operations through convolutional TNNs. Looping back to FL, we propose SWIFT, a novel approach to peer-to-peer (decentralized) federated learning which allows clients to train in an asynchronous, free-flowing manner.
 
Examining Committee

Chair:

Dr. Furong Huang

Department Representative:

Dr. Nirupam Roy

Members:

Dr. Tom Goldstein

Bio

Tahseen Rabbani is a PhD student in the Department of Computer Science at the University of Maryland, College Park. He is advised by Dr. Furong Huang. His research broadly encompasses dimensionality reduction, distributed optimization, and privacy. He is an NSF COMBINE Fellow and Apple Scholars in AI/ML Nominee (UMD).

 

This talk is organized by Tom Hurst