event

PhD Defense by Tyler LaBonte

Primary tabs

Title: Generalization in Dynamic Environments: Foundations and Algorithms

 

Date: Wednesday, April 1st, 2026

Time: 10:00am - 12:00pm EDT

Location: Coda C1215 Midtown

Teams link: Tyler LaBonte Thesis Defense | Meeting-Join | Microsoft Teams

 

Tyler LaBonte

Machine Learning Ph.D. Candidate

School of Industrial and Systems Engineering

Georgia Institute of Technology

 

Committee

1. Dr. Vidya Muthukumar (Advisor), School of Electrical and Computer Engineering and School of Industrial and Systems Engineering, Georgia Institute of Technology 

2. Dr. Jacob Abernethy (Co-advisor), School of Computer Science, Georgia Institute of Technology

3. Dr. Steve Mussmann, School of Computer Science, Georgia Institute of Technology

4. Dr. Justin Romberg, School of Electrical and Computer Engineering, Georgia Institute of Technology

5. Dr. Surbhi Goel, School of Engineering and Applied Science, University of Pennsylvania

 

Abstract

Despite the success of modern machine learning models in a vast number of data-rich domains, an incomplete understanding of their robustness limits deployment in high-consequence scenarios. In particular, generalization often suffers as real-world conditions deviate from training environments and the strict assumptions of classical theory no longer apply. These statistical phenomena precipitate serious impacts in machine learning applications: for example, medical diagnosis algorithms can fail when deployed in new hospitals, and recidivism prediction is known to exacerbate bias against demographic minorities.

 

This thesis represents a joint theoretical and empirical investigation of generalization in dynamic environments. Its core tenet is that an understanding of the underlying mechanisms of generalization phenomena can elicit elegant methodological improvements. Therefore, we place equal emphasis on foundations -- mathematical or experimental analyses which divulge fundamental insights into model behavior -- and algorithms -- adaptive methods which ensure model trustworthiness via rigorous benchmarking or theoretical guarantees.

 

We begin the talk by summarizing proposal work in spurious correlations: patterns which are predictive of the target class in the training dataset but irrelevant to the true classification function, and task shift: wherein the objective function at test-time differs from the training objective. The majority of the talk will concern new work on a theoretical characterization of the spurious feature learning dynamics of a two-layer ReLU neural network in a Boolean data setting. We show that a linear spurious correlation is learned exponentially fast and greatly suppresses the learning of a quadratic core feature. To our knowledge, our results comprise the first end-to-end theoretical analysis of spurious feature learning in neural networks outside the lazy (linearized) regime.

 

Status

  • Workflow status: Published
  • Created by: Tatianna Richardson
  • Created: 03/24/2026
  • Modified By: Tatianna Richardson
  • Modified: 03/24/2026

Categories

Keywords

User Data

Target Audience