<node id="667011">
  <nid>667011</nid>
  <type>event</type>
  <uid>
    <user id="28475"><![CDATA[28475]]></user>
  </uid>
  <created>1680298322</created>
  <changed>1680298322</changed>
  <title><![CDATA[Ph.D. Dissertation Defense - Chao-Han Yang]]></title>
  <body><![CDATA[<p><span><span><strong><span>Title</span></strong><em><span>:&nbsp; </span></em><em><span>A Perturbation Approach to Differential Privacy for Deep Learning based Speech Processing</span></em></span></span></p>

<p><span><span><strong><span>Committee:</span></strong></span></span></p>

<p><span><span><span>Dr. </span><span>Chin-Hui Lee, ECE, Chair</span><span>, Advisor</span></span></span></p>

<p><span><span><span>Dr. </span><span>Larry Heck, ECE</span></span></span></p>

<p><span><span><span>Dr. </span><span>Elliott Moore II, ECE</span></span></span></p>

<p><span><span><span>Dr. </span><span>Pin-Yu Chen, IBM AI</span></span></span></p>

<p><span><span><span>Dr. </span><span>Sabato Marco Siniscalchi, NTNU</span></span></span></p>
]]></body>
  <field_summary_sentence>
    <item>
      <value><![CDATA[A Perturbation Approach to Differential Privacy for Deep Learning based Speech Processing ]]></value>
    </item>
  </field_summary_sentence>
  <field_summary>
    <item>
      <value><![CDATA[<p>To deploy high-performance speech applications, deep neural network (DNN) based models are widely utilized and trained with speech data. New data regulations (e.g., GDPR and CCPA) require service providers to ensure necessary privacy protections through privacy measurements to avoid security threats from query-based malicious attacks. Differential privacy (DP) is a technique that provides identity protection with a mathematical definition for measuring privacy losses under query-based attacks. This dissertation aims to establish a perturbation-based machine learning framework for training with DNNs to maintain the high performance of speech models while satisfying DP requirements. We develop specific mechanisms for signal distortion by adding bounded noise to the speech training data. These perturbations can be analyzed through statistical measurements and estimations of the deployed models. Our theoretical studies have used both Laplace and Gaussian noise to guarantee the privacy budget under a bounded max-divergence measurement with ensemble learning. We further leverage upon the properties of Lipchitz continuity of DNNs to provide model robustness estimation under different perturbations in vector space. Speech recognition in both isolated speech commands and large vocabulary continuous speech settings has been utilized to corroborate our theoretical justifications. We have advanced the proposed framework into popular speech and acoustic modeling tasks and conducted preliminary investigations to explore the trade-off between model performance and DP budgets. We introduce two training frameworks to encounter the model degradation thorough (i) teacher-ensemble learning and (ii) sub-sampling-based training. Our findings show that even with state-of-the-art speaker protection techniques, speaker identity information can still be leaked without DP-based data extraction or training methods. For more applications, we have incorporated decentralized and federated training pipelines of DNNs into the proposed privacy-preserving speech processing mechanism to investigate potential solutions for on-device or cloud-based speech applications for end-users.</p>
]]></value>
    </item>
  </field_summary>
  <field_time>
    <item>
      <value><![CDATA[2023-04-12T12:00:00-04:00]]></value>
      <value2><![CDATA[2023-04-12T14:00:00-04:00]]></value2>
      <rrule><![CDATA[]]></rrule>
      <timezone><![CDATA[America/New_York]]></timezone>
    </item>
  </field_time>
  <field_fee>
    <item>
      <value><![CDATA[]]></value>
    </item>
  </field_fee>
  <field_extras>
      </field_extras>
  <field_audience>
          <item>
        <value><![CDATA[Public]]></value>
      </item>
      </field_audience>
  <field_media>
      </field_media>
  <field_contact>
    <item>
      <value><![CDATA[]]></value>
    </item>
  </field_contact>
  <field_location>
    <item>
      <value><![CDATA[CSIP Library and Teams: https://tinyurl.com/dp-speech]]></value>
    </item>
  </field_location>
  <field_sidebar>
    <item>
      <value><![CDATA[]]></value>
    </item>
  </field_sidebar>
  <field_phone>
    <item>
      <value><![CDATA[]]></value>
    </item>
  </field_phone>
  <field_url>
    <item>
      <url><![CDATA[]]></url>
      <title><![CDATA[]]></title>
            <attributes><![CDATA[]]></attributes>
    </item>
  </field_url>
  <field_email>
    <item>
      <email><![CDATA[]]></email>
    </item>
  </field_email>
  <field_boilerplate>
    <item>
      <nid><![CDATA[]]></nid>
    </item>
  </field_boilerplate>
  <links_related>
      </links_related>
  <files>
      </files>
  <og_groups>
          <item>434381</item>
      </og_groups>
  <og_groups_both>
          <item><![CDATA[ECE Ph.D. Dissertation Defenses]]></item>
      </og_groups_both>
  <field_categories>
          <item>
        <tid>1788</tid>
        <value><![CDATA[Other/Miscellaneous]]></value>
      </item>
      </field_categories>
  <field_keywords>
          <item>
        <tid>192484</tid>
        <value><![CDATA[PhD Defense, graduate students]]></value>
      </item>
      </field_keywords>
  <userdata><![CDATA[]]></userdata>
</node>
