<node id="618695">
  <nid>618695</nid>
  <type>event</type>
  <uid>
    <user id="34963"><![CDATA[34963]]></user>
  </uid>
  <created>1551664376</created>
  <changed>1551664689</changed>
  <title><![CDATA[Modern Statistical Theory Inspired by Deep Learning]]></title>
  <body><![CDATA[<p><strong>Title: </strong>Modern Statistical Theory Inspired by Deep Learning</p>

<p><strong>Abstract:&nbsp;</strong>Modern learning algorithms, such as deep learning, have gained great successes in real applications. However, some&nbsp;of their empirical&nbsp;behaviors&nbsp;may not&nbsp;be interpreted within the&nbsp;classical statistical learning framework. For example, deep learning algorithms achieve small testing error even when the training error is zero, i.e., over-fitting. Another phenomenon is observed in&nbsp;image recognition applications&nbsp;where&nbsp;a hardly noticeable change of data may lead to a dramatic increase&nbsp;in misclassification rates. Inspired by these observations,&nbsp;we attempt&nbsp;to illustrate&nbsp;new theoretical&nbsp;insights for data-interpolation and adversarial testing using the very simple&nbsp;nearest neighbor algorithms. In particular,&nbsp;we prove statistical optimality&nbsp;of interpolated nearest neighbor algorithms. More surprisingly, it is discovered that the classification performance, under a proper interpolation, is even&nbsp;better than the best kNN in terms of multiplicative constant. As for adversarial testing, we demonstrate that different adversarial mechanisms lead to different&nbsp;phase&nbsp;transition phenomena of&nbsp;the misclassification rate in terms of its upper bound. Additionally, our technical&nbsp;analysis invented to deal with adversarial samples&nbsp;can also be applied to other variants&nbsp;of kNN, e.g. pre-processed 1NN and distributed-NN.</p>

<p>&nbsp;</p>

<p><strong>Bio: </strong>Guang Cheng is a Professor of Statistics at Purdue University. &nbsp;He received his Ph.D. in Statistics from the University of Wisconsin-Madison in 2006. &nbsp;His research interests include Big Data and High Dimensional Statistical Inferences, and more recently turned to Deep Learning and Reinforcement Learning. &nbsp;Cheng is the recipient of&nbsp;the NSF CAREER award, Noether Young Scholar Award and Simons Fellowship in Mathematics. Please visit his big data theory research group at&nbsp;<a href="http://www.science.purdue.edu/bigdata/">http://www.science.purdue.edu/bigdata/</a></p>

<p>&nbsp;</p>
]]></body>
  <field_summary_sentence>
    <item>
      <value><![CDATA[Modern Statistical Theory Inspired by Deep Learning]]></value>
    </item>
  </field_summary_sentence>
  <field_summary>
    <item>
      <value><![CDATA[<p><strong>Abstract:&nbsp;</strong>Modern learning algorithms, such as deep learning, have gained great successes in real applications. However, some&nbsp;of their empirical&nbsp;behaviors&nbsp;may not&nbsp;be interpreted within the&nbsp;classical statistical learning framework. For example, deep learning algorithms achieve small testing error even when the training error is zero, i.e., over-fitting. Another phenomenon is observed in&nbsp;image recognition applications&nbsp;where&nbsp;a hardly noticeable change of data may lead to dramatic increase&nbsp;of mis-classification rates. Inspired by these observations,&nbsp;we attempt&nbsp;to illustrate&nbsp;new theoretical&nbsp;insights for data-interpolation and adversarial testing using the very simple&nbsp;nearest neighbor algorithms. In particular,&nbsp;we prove statistical optimality&nbsp;of interpolated nearest neighbor algorithms. More surprisingly, it is discovered that the classification performance, under a proper interpolation, is even&nbsp;better that the best kNN in terms of multiplicative constant. As for adversarial testing, we demonstrate that different adversarial mechanisms lead to different&nbsp;phase&nbsp;transition phenomena of&nbsp;mis-classification rate in terms of its upper bound. Additionally, our technical&nbsp;analysis invented to deal with adversarial samples&nbsp;can also be applied to other variants&nbsp;of kNN, e.g. pre-processed 1NN and distributed-NN.</p>

<p>&nbsp;</p>

<p><strong>Bio: </strong>Guang Cheng is a Professor of Statistics at Purdue University. &nbsp;He received his PhD in Statistics from University of Wisconsin-Madison in 2006. &nbsp;His research interests include Big Data and High Dimensional Statistical Inferences, and more recently turn to Deep Learning and Reinforcement Learning. &nbsp;Cheng is the recipient of&nbsp;the NSF CAREER award, Noether Young Scholar Award and Simons Fellowship in Mathematics. Please visit his big data theory research group at&nbsp;<a href="http://www.science.purdue.edu/bigdata/">http://www.science.purdue.edu/bigdata/</a></p>
]]></value>
    </item>
  </field_summary>
  <field_time>
    <item>
      <value><![CDATA[2019-03-06T13:30:00-05:00]]></value>
      <value2><![CDATA[2019-03-06T14:30:00-05:00]]></value2>
      <rrule><![CDATA[]]></rrule>
      <timezone><![CDATA[America/New_York]]></timezone>
    </item>
  </field_time>
  <field_fee>
    <item>
      <value><![CDATA[]]></value>
    </item>
  </field_fee>
  <field_extras>
      </field_extras>
  <field_audience>
          <item>
        <value><![CDATA[Faculty/Staff]]></value>
      </item>
          <item>
        <value><![CDATA[Postdoc]]></value>
      </item>
          <item>
        <value><![CDATA[Public]]></value>
      </item>
          <item>
        <value><![CDATA[Graduate students]]></value>
      </item>
      </field_audience>
  <field_media>
      </field_media>
  <field_contact>
    <item>
      <value><![CDATA[]]></value>
    </item>
  </field_contact>
  <field_location>
    <item>
      <value><![CDATA[]]></value>
    </item>
  </field_location>
  <field_sidebar>
    <item>
      <value><![CDATA[]]></value>
    </item>
  </field_sidebar>
  <field_phone>
    <item>
      <value><![CDATA[]]></value>
    </item>
  </field_phone>
  <field_url>
    <item>
      <url><![CDATA[https://isye.gatech.edu/about/maps-directions/isye-building-complex]]></url>
      <title><![CDATA[Groseclose Building]]></title>
            <attributes><![CDATA[]]></attributes>
    </item>
  </field_url>
  <field_email>
    <item>
      <email><![CDATA[]]></email>
    </item>
  </field_email>
  <field_boilerplate>
    <item>
      <nid><![CDATA[]]></nid>
    </item>
  </field_boilerplate>
  <links_related>
      </links_related>
  <files>
      </files>
  <og_groups>
          <item>602673</item>
      </og_groups>
  <og_groups_both>
          <item><![CDATA[TRIAD ]]></item>
      </og_groups_both>
  <field_categories>
          <item>
        <tid>1795</tid>
        <value><![CDATA[Seminar/Lecture/Colloquium]]></value>
      </item>
      </field_categories>
  <field_keywords>
          <item>
        <tid>109581</tid>
        <value><![CDATA[deep learning]]></value>
      </item>
      </field_keywords>
  <userdata><![CDATA[]]></userdata>
</node>
