Online Program Home
  My Program

All Times EDT

Abstract Details

Activity Number: 538 - Emerging Topics in Private Data Analysis
Type: Topic Contributed
Date/Time: Thursday, August 6, 2020 : 1:00 PM to 2:50 PM
Sponsor: IMS
Abstract #309724
Title: KNG: The K-Norm Gradient Mechanism
Author(s): Jordan Awan* and Matthew Reimherr
Companies: Penn State University and Penn State University
Keywords: empirical risk minimization; regression; differential privacy

This paper presents a new mechanism for producing sanitized statistical summaries that achieve differential privacy, called the K-Norm Gradient Mechanism, or KNG. This new approach maintains the strong flexibility of the exponential mechanism, while achieving the powerful utility performance of objective perturbation. KNG starts with an inherent objective function (often an empirical risk), and promotes summaries that are close to minimizing the objective by weighting according to how far the gradient of the objective function is from zero. Working with the gradient instead of the original objective function allows for additional flexibility as one can penalize using different norms. We show that, unlike the exponential mechanism, the noise added by KNG is asymptotically negligible compared to the statistical error for many problems. In addition to theoretical guarantees on privacy and utility, we confirm the utility of KNG empirically in the settings of linear and quantile regression through simulations.

Authors who are presenting talks have a * after their name.

Back to the full JSM 2020 program