event

PhD Defense by Kartik Sharma

Primary tabs

Title: Adapting Models to User Intent for Safe and Controllable Generation

 

Date: Tuesday, April 21st , 2026

Time: 2:00 p.m. - 4:00 p.m. ET

Location (in-person + virtual): Coda C1303 Glenwood [Virtual Link]

 

Kartik Sharma

Ph.D. candidate

School of Computational Science and Engineering

College of Computing

Georgia Institute of Technology 

 

Committee:

Dr. Srijan Kumar (Advisor, Computational Science and Engineering, Georgia Institute of Technology)

Dr. Rakshit Trivedi (Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology)

Dr. Chao Zhang (Computational Science and Engineering, Georgia Institute of Technology)

Dr. Bo Dai (Computational Science and Engineering, Georgia Institute of Technology)

Dr. Polo Chao (Computational Science and Engineering, Georgia Institute of Technology)

 

 

Abstract:

As artificial intelligence (AI) is increasingly deployed in high-stakes, open-world environments, the demand for systems that are both highly capable and strictly trustworthy has become paramount. Modern generative models require high test-time plasticity to remain responsive to complex and dynamic human intentions. However, this same architectural flexibility inadvertently leaves them vulnerable to manipulation by bad actors and unintended changes. To ensure AI remains an empowering tool rather than a fragile liability, this dissertation presents a comprehensive framework for adapting models safely to user intent. First, this work introduces interpretable test-time mechanisms that empower users to precisely steer model behavior. These interventions are designed to enforce hard constraints in graph diffusion models and guide large language and imitation learning models toward exemplary and directed behavior descriptions. However, because unconstrained plasticity inherently alters a model’s risk profile, the research then systematically diagnoses the vulnerabilities in the systems, exposing vulnerabilities in dynamic graph models and multimodal language models. To safely support user-driven controllability without succumbing to these adversarial risks, the final phase demonstrates how structural adaptivity of system prompts can protect a language model’s core without compromising its benign behavior. Ultimately, this research works towards resolving the modern stability-plasticity dilemma, transforming vulnerable networks into deeply steerable and secure extensions of human intent.

 

Status

  • Workflow status: Published
  • Created by: Tatianna Richardson
  • Created: 04/03/2026
  • Modified By: Tatianna Richardson
  • Modified: 04/03/2026

Categories

Keywords

User Data

Target Audience