log in  |  register  |  feedback?  |  help  |  web accessibility
PhD Proposal: DEEP LEARNING UNDER REAL-WORLD CONSTRAINTS: EFFICIENCY, STRUCTURE, AND ROBUSTNESS IN THE MODERN ML STACK
Alexander Stein
IRB-4107 https://umd.zoom.us/j/5209864628?pwd=eU5kaEErcnVGZW1PdUdwWWgrenI0Zz09&omn=99225074436&jst=2
Tuesday, June 10, 2025, 2:30-4:00 pm
  • You are subscribed to this talk through .
  • You are watching this talk through .
  • You are subscribed to this talk. (unsubscribe, watch)
  • You are watching this talk. (unwatch, subscribe)
  • You are not subscribed to this talk. (watch, subscribe)
Abstract

This proposal outlines a research agenda focused on improving the performance, efficiency, and robustness of modern deep learning models (with an emphasis on transformer-based neural networks) under real-world constraints. As large language models (LLMs) become increasingly central to AI systems, it is vital to understand how these models behave and perform when deployed outside idealized settings. My work explores this space through three complementary threads: (1) structure-aware modeling in constrained domains like tabular data, where models like STEP demonstrate that simplicity and domain alignment can outperform more complex pipelines; (2) inference-time efficiency, including ongoing work on tokenizer compression and long-context extensions to reduce decoding costs; and (3) robustness and privacy under adversarial pressure, including studies of coercion attacks and information leakage. Together, these projects reflect a cohesive research direction grounded in practicality and broad applicability. Future work will investigate underexplored deployment scenarios, such as tokenization, tabular modeling, and computational trade-offs, with the goal of designing adaptable, efficient, and robust transformer systems.

Bio

Alex is a PhD candidate in Computer Science at the University of Maryland, College Park, advised by Professors Tom Goldstein and John Dickerson. His research interests cover a range of machine learning topics, including but not limited to: LLM reasoning, adversarial attacks, and efficient implementation of transformer architectures. He is especially interested in exploring the extent to which LLMs understand context, reason strategically, and generalize beyond their training data.

During the summer of 2024, he was an Applied Research intern at Capital One as part of the Behavior Modeling team, where he explored using transformers for event prediction. This summer, Alex will be a quantitative researcher on the structured learning team at Two Sigma.

Before joining UMD, Alex was a Vice President in Algorithmic Research and Development at RBC Capital Markets in New York City, where he focused on enhancing the intelligence and infrastructure behind the Equities division’s proprietary execution platform.

Alex holds a Bachelors of Science in Computer Science and Operations Research from Columbia University.

This talk is organized by Migo Gui