<node id="671172">
  <nid>671172</nid>
  <type>event</type>
  <uid>
    <user id="27707"><![CDATA[27707]]></user>
  </uid>
  <created>1700518226</created>
  <changed>1700518410</changed>
  <title><![CDATA[PhD Proposal by Shreyas Malakarjun Patil]]></title>
  <body><![CDATA[<p><span><span><strong><span><span><span><span>Title: </span></span></span></span></strong><span><span><span><span>Leveraging sparsity in deep neural networks for training efficiency, interpretability and transfer learning</span></span></span></span></span></span></p>

<p><span><span><span>&nbsp;</span></span></span></p>

<p><span><span><span><strong><span><span><span>Date: </span></span></span></strong><span><span><span>November 30th&nbsp;</span></span></span></span></span></span></p>

<p><span><span><span><strong><span><span><span>Time: </span></span></span></strong><span><span><span>10:30 AM</span></span></span></span></span></span></p>

<p><span><span><span><strong><span><span><span>Physical attendance</span></span></span></strong><span><span><span>: conference room (Midtown) on Coda 12th&nbsp;floor</span></span></span></span></span></span></p>

<p><span><span><span><strong><span><span><span>Virtual <span>attendance</span></span></span></span></strong><span><span><span>: <span><a href="https://gatech.zoom.us/j/96615684093" target="_blank">https://gatech.zoom.us/j/96615684093</a></span></span></span></span></span></span></span></p>

<p>&nbsp;</p>

<p><span><span><span><strong><span><span><span>Shreyas Malakarjun Patil</span></span></span></strong></span></span></span></p>

<p><span><span><span><span><span><span>Machine Learning PhD Student</span></span></span></span></span></span></p>

<p><span><span><span><span><span><span>ECE<br />
Georgia Institute of Technology</span></span></span></span></span></span></p>

<p><span><span><span>&nbsp;</span></span></span></p>

<p><span><span><span><strong><span><span><span>Committee</span></span></span></strong></span></span></span></p>

<p><span><span><span><span><span><span>1. Dr. Constantine Dovrolis (Advisor)</span></span></span></span></span></span></p>

<p><span><span><span><span><span><span>2. Dr. Ling Liu</span></span></span></span></span></span></p>

<p><span><span><span><span><span><span>3. Dr. Zsolt Kira</span></span></span></span></span></span></p>

<p><span><span><span>&nbsp;</span></span></span></p>

<p><span><span><span><strong><span><span><span>Abstract</span></span></span></strong></span></span></span></p>

<p>&nbsp;</p>

<p><span><span><span><span><span><span>Sparse neural networks (NNs) exhibit fewer connections between consecutive layers compared to dense NNs. As a result, sparse NNs have been shown to enhance generalization and computational efficiency. However, the diverse sparse network structures and benefits beyond efficiency and generalization remain largely unexplored.&nbsp;</span></span></span></span></span></span></p>

<p>&nbsp;</p>

<p><span><span><span><span><span><span>In this dissertation, I present an exploration of sparse network structures and their ensuing benefits. First, we propose a new method, PHEW, to identify sparse NNs at initialization without using training data. PHEW leads to sparse NNs that learn fast and generalize well, thus enhancing training efficiency. Second, we propose Neural Sculpting to uncover the hierarchically modular task structure in NNs. We iteratively prune units and edges during training and combine it with network analysis to detect modules and infer hierarchy, thereby enhancing NN interpretability. Finally, we plan to examine how efficiently hierarchically modular NNs, that reflect the task’s structure, transfer to new tasks as compared to dense NNs. Given the assumption that the new tasks introduced in transfer learning share similarities with the previous tasks, our investigation will specifically explore the degree of sub-task reuse from the initial tasks. In summary, this dissertation advances the understanding and capabilities of sparse NNs in terms of training efficiency, interpretability, and transfer learning.</span></span></span></span></span></span></p>
]]></body>
  <field_summary_sentence>
    <item>
      <value><![CDATA[Leveraging sparsity in deep neural networks for training efficiency, interpretability and transfer learning]]></value>
    </item>
  </field_summary_sentence>
  <field_summary>
    <item>
      <value><![CDATA[<p><span><span><span><span>Leveraging sparsity in deep neural networks for training efficiency, interpretability and transfer learning</span></span></span></span></p>
]]></value>
    </item>
  </field_summary>
  <field_time>
    <item>
      <value><![CDATA[2023-11-30T10:30:00-05:00]]></value>
      <value2><![CDATA[2023-11-30T12:00:00-05:00]]></value2>
      <rrule><![CDATA[]]></rrule>
      <timezone><![CDATA[America/New_York]]></timezone>
    </item>
  </field_time>
  <field_fee>
    <item>
      <value><![CDATA[]]></value>
    </item>
  </field_fee>
  <field_extras>
      </field_extras>
  <field_audience>
          <item>
        <value><![CDATA[Public]]></value>
      </item>
      </field_audience>
  <field_media>
      </field_media>
  <field_contact>
    <item>
      <value><![CDATA[]]></value>
    </item>
  </field_contact>
  <field_location>
    <item>
      <value><![CDATA[conference room (Midtown) on Coda 12th floor]]></value>
    </item>
  </field_location>
  <field_sidebar>
    <item>
      <value><![CDATA[]]></value>
    </item>
  </field_sidebar>
  <field_phone>
    <item>
      <value><![CDATA[]]></value>
    </item>
  </field_phone>
  <field_url>
    <item>
      <url><![CDATA[]]></url>
      <title><![CDATA[]]></title>
            <attributes><![CDATA[]]></attributes>
    </item>
  </field_url>
  <field_email>
    <item>
      <email><![CDATA[]]></email>
    </item>
  </field_email>
  <field_boilerplate>
    <item>
      <nid><![CDATA[]]></nid>
    </item>
  </field_boilerplate>
  <links_related>
      </links_related>
  <files>
      </files>
  <og_groups>
          <item>221981</item>
      </og_groups>
  <og_groups_both>
          <item><![CDATA[Graduate Studies]]></item>
      </og_groups_both>
  <field_categories>
          <item>
        <tid>1788</tid>
        <value><![CDATA[Other/Miscellaneous]]></value>
      </item>
      </field_categories>
  <field_keywords>
          <item>
        <tid>102851</tid>
        <value><![CDATA[Phd proposal]]></value>
      </item>
      </field_keywords>
  <userdata><![CDATA[]]></userdata>
</node>
