<node id="685253">
  <nid>685253</nid>
  <type>event</type>
  <uid>
    <user id="28475"><![CDATA[28475]]></user>
  </uid>
  <created>1758748056</created>
  <changed>1758748182</changed>
  <title><![CDATA[Ph.D. Proposal Oral Exam - Abhimanyu Bambhaniya]]></title>
  <body><![CDATA[<p><strong>Title:&nbsp; </strong><em>Algorithmic Optimizations and Distributed Platform Design for LLM Inference</em></p><p><strong>Committee:&nbsp;</strong></p><p>Dr.&nbsp;Krishna, Advisor&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;</p><p>Dr. Tumanov, Chair</p><p>Dr. Raychowdhury</p><p>Dr. Subramanian&nbsp;</p>]]></body>
  <field_summary_sentence>
    <item>
      <value><![CDATA[Algorithmic Optimizations and Distributed Platform Design for LLM Inference]]></value>
    </item>
  </field_summary_sentence>
  <field_summary>
    <item>
      <value><![CDATA[<p>The objective of the proposed research is to develop systematic approaches to Next-Generation AI inference platform design that address the complex optimization challenges spanning multiple interdependent layers of modern LLM serving infrastructure. Large Language Model (LLM) inference platforms have become critical for deploying AI at scale, yet their design involves navigating an enormous space of architectural choices, from evolving model architectures and optimization techniques to heterogeneous hardware platforms. Current approaches rely heavily on empirical testing of limited configurations, lacking comprehensive frameworks for evaluating the interplay between workload characteristics, system optimizations, and hardware architectures. This disconnect between algorithmic innovations and hardware capabilities results in suboptimal deployments, inefficient resource utilization, and missed opportunities for performance improvements across the increasingly complex multi-stage inference pipelines that characterize real-world AI applications. This thesis proposes a systematic approach to AI inference platform design through comprehensive modeling and simulation frameworks. GenZ, an analytical framework, enables rapid exploration of LLM architectures and optimizations across diverse current and next-generation NPU platforms, providing insights for future hardware design. MIST extends this capability to full end-to-end inference pipelines, modeling complex multi-client scenarios with heterogeneous batching strategies and hardware configurations that existing simulators cannot capture. Additionally, it studies hardware-software co-design for sparse attention accelerators, demonstrating systematic methodologies for architectural optimization under sparsity constraints. Finally, we propose training recipes for structured N:M sparsity in transformer models, achieving up to 5% accuracy improvements in high-sparsity regimes through progressive gradient flow techniques. The thesis also proposes MoE optimization strategies for multi-host inference and next-generation platforms for agentic AI workloads.</p>]]></value>
    </item>
  </field_summary>
  <field_time>
    <item>
      <value><![CDATA[2025-09-30T12:00:00-04:00]]></value>
      <value2><![CDATA[2025-09-30T14:00:00-04:00]]></value2>
      <rrule><![CDATA[]]></rrule>
      <timezone><![CDATA[America/New_York]]></timezone>
    </item>
  </field_time>
  <field_fee>
    <item>
      <value><![CDATA[]]></value>
    </item>
  </field_fee>
  <field_extras>
      </field_extras>
  <field_audience>
          <item>
        <value><![CDATA[Public]]></value>
      </item>
      </field_audience>
  <field_media>
      </field_media>
  <field_contact>
    <item>
      <value><![CDATA[]]></value>
    </item>
  </field_contact>
  <field_location>
    <item>
      <value><![CDATA[Room 3126, Klaus ]]></value>
    </item>
  </field_location>
  <field_sidebar>
    <item>
      <value><![CDATA[]]></value>
    </item>
  </field_sidebar>
  <field_phone>
    <item>
      <value><![CDATA[]]></value>
    </item>
  </field_phone>
  <field_url>
    <item>
      <url><![CDATA[]]></url>
      <title><![CDATA[]]></title>
            <attributes><![CDATA[]]></attributes>
    </item>
  </field_url>
  <field_email>
    <item>
      <email><![CDATA[]]></email>
    </item>
  </field_email>
  <field_boilerplate>
    <item>
      <nid><![CDATA[]]></nid>
    </item>
  </field_boilerplate>
  <links_related>
          <item>
        <url>https://gatech.zoom.us/j/97053430439?pwd=gUNSlCe3phYeJMuc0FmOKvRc1ZBXAw.1</url>
        <link_title><![CDATA[Zoom link]]></link_title>
      </item>
      </links_related>
  <files>
      </files>
  <og_groups>
          <item>434371</item>
      </og_groups>
  <og_groups_both>
          <item><![CDATA[ECE Ph.D. Proposal Oral Exams]]></item>
      </og_groups_both>
  <field_categories>
          <item>
        <tid>1788</tid>
        <value><![CDATA[Other/Miscellaneous]]></value>
      </item>
      </field_categories>
  <field_keywords>
          <item>
        <tid>102851</tid>
        <value><![CDATA[Phd proposal]]></value>
      </item>
          <item>
        <tid>1808</tid>
        <value><![CDATA[graduate students]]></value>
      </item>
      </field_keywords>
  <field_userdata><![CDATA[]]></field_userdata>
</node>
