<node id="674896">
  <nid>674896</nid>
  <type>event</type>
  <uid>
    <user id="27707"><![CDATA[27707]]></user>
  </uid>
  <created>1716579293</created>
  <changed>1716579318</changed>
  <title><![CDATA[PhD Defense by Jiaao Chen]]></title>
  <body><![CDATA[<p><strong>Title:</strong>&nbsp; Efficient and Adaptive Machine Learning for Natural Language Processing</p><p>&nbsp;</p><p><strong>Date:</strong> Tuesday, June 4, 2024</p><p><strong>Time:</strong> 11:00am - 1:30pm EST</p><p><strong>Location (Zoom)</strong>:&nbsp;<a href="https://gatech.zoom.us/j/93873694265?pwd=RUNaNWJLclhxMzBQM3BGNmJobVZrZz09" title="https://gatech.zoom.us/j/93873694265?pwd=RUNaNWJLclhxMzBQM3BGNmJobVZrZz09">https://gatech.zoom.us/j/93873694265?pwd=RUNaNWJLclhxMzBQM3BGNmJobVZrZz09</a>&nbsp;</p><p>&nbsp;</p><p><a href="https://cs.stanford.edu/people/jiaaoc/" title="https://cs.stanford.edu/people/jiaaoc/"><strong>Jiaao Chen</strong></a></p><p>PhD Candidate</p><p>School of Interactive Computing</p><p>College of Computing</p><p>Georgia Institute of Technology&nbsp;</p><p>&nbsp;</p><p><strong>Committee</strong>:</p><p>Dr. Diyi Yang (advisor), Computer Science Department, Stanford University</p><p>Dr. Mark Riedl (co-advisor), School of Interactive Computing, Georgia Tech</p><p>Dr. Alan Ritter, School of Interactive Computing, Georgia Tech</p><p>Dr. Zsolt Kira, School of Interactive Computing, Georgia Tech</p><p>Dr. Colin Raffel, Department of Computer Science, University of Toronto</p><p>&nbsp;</p><p><strong>Abstract:</strong></p><p>NLP has recently undergone a transformational shift towards the development and application of Large Language Models (LLMs), which represent a significant leap in performance. However, the extensive computational resources and the vast textual data sets on which they rely create key challenges in scenarios with limited resources, such as limited data, computational power, memory, and specialized expertise. I argue that developing machine learning methods that could efficiently adapt with limited resources is particularly vital in the era of LLMs to make the benefits of the technology sustainable, accessible, and generalizable.</p><p>&nbsp;</p><p>In this thesis, I advocate for <strong>efficient and adaptive machine learning for NLP</strong>, endeavoring to make NLP models benefit real-world applications by diving into three parts. Part I tackles the essential element of NLP learning, data. I propose to improve the data efficiency for developing NLP systems by&nbsp;augmenting data through hidden space manipulation, linguistically informed perturbation, human learning strategies based generation and reasoning graphs. Part II deals with the information beyond the text in languages, structures. I explore what and how structures could be explicitly incorporated to enhance the NLP models. Part III focuses on the adapting process of NLP models, training. I investigate how we can efficiently fine-tune, update, and unlearn NLP models.</p><p><br>From improving data efficiency and incorporating structures to improving training efficiency, this thesis aims to <strong>develop efficient and adaptive machine learning for NLP to democratize NLP systems with scarce resources, specialized fields, and emerging applications</strong>.</p>]]></body>
  <field_summary_sentence>
    <item>
      <value><![CDATA[Efficient and Adaptive Machine Learning for Natural Language Processing]]></value>
    </item>
  </field_summary_sentence>
  <field_summary>
    <item>
      <value><![CDATA[<p>Efficient and Adaptive Machine Learning for Natural Language Processing</p>]]></value>
    </item>
  </field_summary>
  <field_time>
    <item>
      <value><![CDATA[2024-06-04T11:00:00-04:00]]></value>
      <value2><![CDATA[2024-06-04T13:00:00-04:00]]></value2>
      <rrule><![CDATA[]]></rrule>
      <timezone><![CDATA[America/New_York]]></timezone>
    </item>
  </field_time>
  <field_fee>
    <item>
      <value><![CDATA[]]></value>
    </item>
  </field_fee>
  <field_extras>
      </field_extras>
  <field_audience>
          <item>
        <value><![CDATA[Public]]></value>
      </item>
      </field_audience>
  <field_media>
      </field_media>
  <field_contact>
    <item>
      <value><![CDATA[]]></value>
    </item>
  </field_contact>
  <field_location>
    <item>
      <value><![CDATA[Zoom]]></value>
    </item>
  </field_location>
  <field_sidebar>
    <item>
      <value><![CDATA[]]></value>
    </item>
  </field_sidebar>
  <field_phone>
    <item>
      <value><![CDATA[]]></value>
    </item>
  </field_phone>
  <field_url>
    <item>
      <url><![CDATA[]]></url>
      <title><![CDATA[]]></title>
            <attributes><![CDATA[]]></attributes>
    </item>
  </field_url>
  <field_email>
    <item>
      <email><![CDATA[]]></email>
    </item>
  </field_email>
  <field_boilerplate>
    <item>
      <nid><![CDATA[]]></nid>
    </item>
  </field_boilerplate>
  <links_related>
      </links_related>
  <files>
      </files>
  <og_groups>
          <item>221981</item>
      </og_groups>
  <og_groups_both>
          <item><![CDATA[Graduate Studies]]></item>
      </og_groups_both>
  <field_categories>
          <item>
        <tid>1788</tid>
        <value><![CDATA[Other/Miscellaneous]]></value>
      </item>
      </field_categories>
  <field_keywords>
          <item>
        <tid>100811</tid>
        <value><![CDATA[Phd Defense]]></value>
      </item>
      </field_keywords>
  <field_userdata><![CDATA[]]></field_userdata>
</node>
