<node id="674612">
  <nid>674612</nid>
  <type>event</type>
  <uid>
    <user id="28475"><![CDATA[28475]]></user>
  </uid>
  <created>1715244552</created>
  <changed>1715244653</changed>
  <title><![CDATA[Ph.D. Proposal Oral Exam - Woohong Byun]]></title>
  <body><![CDATA[<p><strong>Title:&nbsp; </strong><em>Energy-efficient Hardware Acceleration of Transformer-based Models</em></p>

<p><strong>Committee:&nbsp; </strong></p>

<p>Dr. Mukhopadhyay, Advisor&nbsp;&nbsp;&nbsp;</p>

<p>Dr. Yu, Chair</p>

<p>Dr. Sathe</p>
]]></body>
  <field_summary_sentence>
    <item>
      <value><![CDATA[Energy-efficient Hardware Acceleration of Transformer-based Models]]></value>
    </item>
  </field_summary_sentence>
  <field_summary>
    <item>
      <value><![CDATA[<p>The objective of the proposed research is to develop an energy-efficient hardware accelerator for transformer-based models. Transformer-based language models have demonstrated remarkable performance in various natural language processing (NLP) tasks. However, their substantial memory requirements pose not only the challenge of enabling these models on edge devices but also the issue of managing their energy consumption on servers. To address these challenges, this study introduces a novel parameter quantization method for BERT. This method utilizes a Hessian-based sensitivity metric to allocate higher precision bit widths to important parameters and lower precision bit widths to less critical ones. Expanding on this method, the research proposes a new energy-efficient software-hardware co-optimization approach. This approach encompasses a hardware-friendly hessian-based row-wise mixed-precision quantization algorithm and an FPGA-based accelerator, allowing an FPGA to accommodate all necessary parameters of the BERT base model without the need for off-chip memory access during runtime. Future research efforts will extend to the development of effective quantization methods tailored for even larger and more complex language models, such as GPT-3 and OPT. Furthermore, the study will focus on designing energy-efficient accelerators specifically customized for these larger language models.</p>
]]></value>
    </item>
  </field_summary>
  <field_time>
    <item>
      <value><![CDATA[2024-05-13T13:00:00-04:00]]></value>
      <value2><![CDATA[2024-05-13T15:00:00-04:00]]></value2>
      <rrule><![CDATA[]]></rrule>
      <timezone><![CDATA[America/New_York]]></timezone>
    </item>
  </field_time>
  <field_fee>
    <item>
      <value><![CDATA[]]></value>
    </item>
  </field_fee>
  <field_extras>
      </field_extras>
  <field_audience>
          <item>
        <value><![CDATA[Public]]></value>
      </item>
      </field_audience>
  <field_media>
      </field_media>
  <field_contact>
    <item>
      <value><![CDATA[]]></value>
    </item>
  </field_contact>
  <field_location>
    <item>
      <value><![CDATA[Online]]></value>
    </item>
  </field_location>
  <field_sidebar>
    <item>
      <value><![CDATA[]]></value>
    </item>
  </field_sidebar>
  <field_phone>
    <item>
      <value><![CDATA[]]></value>
    </item>
  </field_phone>
  <field_url>
    <item>
      <url><![CDATA[]]></url>
      <title><![CDATA[]]></title>
            <attributes><![CDATA[]]></attributes>
    </item>
  </field_url>
  <field_email>
    <item>
      <email><![CDATA[]]></email>
    </item>
  </field_email>
  <field_boilerplate>
    <item>
      <nid><![CDATA[]]></nid>
    </item>
  </field_boilerplate>
  <links_related>
          <item>
        <url>https://teams.microsoft.com/l/meetup-join/19%3ameeting_Yzc2MGU4ZGYtMzNkZC00ODQ4LTlmMzctYTE5NDlmYmQxNTE0%40thread.v2/0?context=%7b%22Tid%22%3a%22482198bb-ae7b-4b25-8b7a-6d7f32faa083%22%2c%22Oid%22%3a%224f74ada8-7c29-4bba-a4ad-2cf7214f2aa0%22%7d</url>
        <link_title><![CDATA[Microsoft Teams Meeting link]]></link_title>
      </item>
      </links_related>
  <files>
      </files>
  <og_groups>
          <item>434371</item>
      </og_groups>
  <og_groups_both>
          <item><![CDATA[ECE Ph.D. Proposal Oral Exams]]></item>
      </og_groups_both>
  <field_categories>
          <item>
        <tid>1788</tid>
        <value><![CDATA[Other/Miscellaneous]]></value>
      </item>
      </field_categories>
  <field_keywords>
          <item>
        <tid>102851</tid>
        <value><![CDATA[Phd proposal]]></value>
      </item>
          <item>
        <tid>1808</tid>
        <value><![CDATA[graduate students]]></value>
      </item>
      </field_keywords>
  <field_userdata><![CDATA[]]></field_userdata>
</node>
