{"681539":{"#nid":"681539","#data":{"type":"event","title":"PhD Defense | On the Resource Efficiency of Language Models","body":[{"value":"\u003Cp\u003E\u003Cstrong\u003ETitle: On the Resource Efficiency of Language Models\u003C\/strong\u003E\u003C\/p\u003E\u003Cp\u003E\u0026nbsp;\u003C\/p\u003E\u003Cp\u003E\u003Cstrong\u003EDate:\u0026nbsp;\u003C\/strong\u003EApril 16th, 2025\u003C\/p\u003E\u003Cp\u003E\u003Cstrong\u003ETime:\u0026nbsp;\u003C\/strong\u003E1:00 pm - 3:00 pm (EST)\u003C\/p\u003E\u003Cp\u003ELocation: CODA C1308\u003C\/p\u003E\u003Cp\u003EMeeting URL: \u003Ca href=\u0022https:\/\/gatech.zoom.us\/j\/91275576046\u0022\u003Ehttps:\/\/gatech.zoom.us\/j\/91275576046\u003C\/a\u003E\u0026nbsp;\u003C\/p\u003E\u003Cp\u003EMeeting ID: 912 7557 6046\u003C\/p\u003E\u003Cp\u003E\u0026nbsp;\u003C\/p\u003E\u003Cp\u003E\u003Cstrong\u003ERongzhi Zhang\u003C\/strong\u003E\u003C\/p\u003E\u003Cp\u003EMachine Learning PhD Candidate\u003C\/p\u003E\u003Cp\u003ESchool of Computational Science and Engineering\u003Cbr\u003EGeorgia Institute of Technology\u003C\/p\u003E\u003Cp\u003E\u0026nbsp;\u003C\/p\u003E\u003Cp\u003E\u003Cstrong\u003ECommittee\u003C\/strong\u003E\u003C\/p\u003E\u003Cp\u003E1. Dr. Chao Zhang\u0026nbsp;(CSE, Georgia Tech) (Advisor)\u003C\/p\u003E\u003Cp\u003E2. Dr. Tuo Zhao (ISyE, Georgia Tech)\u003C\/p\u003E\u003Cp\u003E3. Dr. Steve Mussmann (CS, Georgia Tech)\u003C\/p\u003E\u003Cp\u003E4. Dr. B. Aditya Prakash \u0026nbsp;(CSE, Georgia Tech)\u0026nbsp;\u003C\/p\u003E\u003Cp\u003E5. Dr. Yelong Shen (Microsoft)\u003C\/p\u003E\u003Cp\u003E\u0026nbsp;\u003C\/p\u003E\u003Cp\u003E\u003Cstrong\u003EAbstract\u003C\/strong\u003E\u003Cbr\u003ELarge Language Models (LLMs) have achieved remarkable progress across natural language processing tasks, yet their broad application remains constrained by resource challenges. This thesis addresses these challenges from two complementary thrusts: data efficiency in the post-training stage and model efficiency in the deployment stage. The proposed approaches aim to reduce supervision and memory requirements while preserving or even enhancing downstream performance.\u003C\/p\u003E\u003Cp\u003E\u0026nbsp;\u003C\/p\u003E\u003Cp\u003E\u003Cstrong\u003EThrust I: Data Efficiency in the Post-Training Stage\u003C\/strong\u003E\u003C\/p\u003E\u003Cp\u003EIn the post-training stage, adaptation to specific tasks or alignment with human values demands large quantities of high-quality labeled data. To improve data curation efficiency for fine-tuning pre-trained language models, I introduce PRBoost, an interactive weak supervision framework that iteratively discovers labeling rules, which mitigates the data scarcity issue and boosts model performance over existing weakly-supervised baselines. To improve data utilization efficiency in LLM alignment, I propose DORM, a two-stage approach that dynamically adjusts preference data weights via quality-aware weighting and bilevel optimization, achieving strong alignment results using up to 40\u00d7 less data than conventional techniques.\u003C\/p\u003E\u003Cp\u003E\u0026nbsp;\u003C\/p\u003E\u003Cp\u003E\u003Cstrong\u003EThrust II: Model Efficiency in the Deployment Stage\u003C\/strong\u003E\u003C\/p\u003E\u003Cp\u003EIn the deployment stage, the use of LLMs in resource-limited environments is constrained by their enormous parameter counts and memory requirements. To enhance model parameter efficiency, I develop PTLoss, a perturbation-based distillation framework that improves student model performance when distilling from biased teacher models. To enhance model efficiency during inference, I present LoRC, a progressive KV cache compression strategy based on low-rank approximations of KV weight matrices, which demonstrates substantial GPU memory savings with minimal performance degradation.\u003Cbr\u003E\u003Cbr\u003ETogether, these contributions establish a comprehensive framework for resource-efficient language models, enabling more practical application of LLMs across resource-constrained environments.\u003Cbr\u003E\u003Cbr\u003E\u0026nbsp;\u003C\/p\u003E","summary":"","format":"limited_html"}],"field_subtitle":"","field_summary":[{"value":"\u003Cp\u003E\u003Cstrong\u003EOn the Resource Efficiency of Language Models\u003C\/strong\u003E\u003C\/p\u003E","format":"limited_html"}],"field_summary_sentence":[{"value":"Rongzhi Zhang - Machine Learning PhD Student - School of Computational Science and Engineering"}],"uid":"36518","created_gmt":"2025-04-03 13:23:32","changed_gmt":"2025-04-03 13:24:14","author":"shatcher8","boilerplate_text":"","field_publication":"","field_article_url":"","field_event_time":{"event_time_start":"2025-04-16T13:00:00-04:00","event_time_end":"2025-04-16T15:00:00-04:00","event_time_end_last":"2025-04-16T15:00:00-04:00","gmt_time_start":"2025-04-16 17:00:00","gmt_time_end":"2025-04-16 19:00:00","gmt_time_end_last":"2025-04-16 19:00:00","rrule":null,"timezone":"America\/New_York"},"location":"CODA C1308","extras":[],"groups":[{"id":"576481","name":"ML@GT"}],"categories":[],"keywords":[],"core_research_areas":[],"news_room_topics":[],"event_categories":[],"invited_audience":[],"affiliations":[],"classification":[],"areas_of_expertise":[],"news_and_recent_appearances":[],"phone":[],"contact":[],"email":[],"slides":[],"orientation":[],"userdata":""}}}