高级检索

    数据治理中的隐私审计研究

    Research on Privacy Auditing in Data Governance

    • 摘要: 隐私审计是数据治理中的关键问题,旨在判断数据的隐私是否得到了有效保护. 通常,学者们通过对数据添加噪声或扰动实现差分隐私,从而保护个人隐私. 特别在机器学习场景下,出现越来越多的差分隐私算法, 并且这些算法均声称自己可以达到较为严格的隐私保护水平. 然而,即使这些算法在发布之前会经过严格的数学证明,其实际应用中的隐私保护程度亦难以确定. 鉴于差分隐私理论本身的复杂性,隐私算法中证明的错误和编程实现的错误时有发生,使得这些算法无法达到其声称的隐私保护水平,导致隐私泄露. 为了解决这一问题,隐私审计应运而生. 隐私审计可以获取隐私算法的真实隐私保护水平,有助于算法设计者对算法进行改进. 将综述隐私审计相关算法,从数据构建、数据测算、结果量化3个维度进行总结,并对隐私审计算法进行实验说明,最终提出隐私审计面临的挑战以及未来研究方向.

       

      Abstract: Privacy auditing is a crucial issue of data governance, aiming to detect whether data privacy has been protected effectively. Typically, scholars would protect personal private data to meet differential privacy guarantees by perturbing data or adding noise to them. Especially in scenarios of machine learning, an increasing number of differential privacy algorithms have emerged, claiming a relatively stringent level of privacy protection. Although rigorous mathematical proofs of privacy have been conducted before the algorithms’ release, the actual effect on privacy in practice is hardly assured. Due to the complexity of the theory of differential privacy, the correctness of their proofs may not have been thoroughly examined, and imperceptible errors may occur during programming. All of these can undermine the extent of privacy protection to the claimed degree, leaking additional privacy. To tackle this issue, privacy auditing for differential privacy algorithms has emerged. This technique aims to obtain the actual degree of privacy-preserving of differential privacy algorithms, facilitating the discovery of mistakes and improving existing differential privacy algorithms. This paper surveys the scenarios and methods of privacy auditing, summarizing the methods from three aspects―data construction, data measurement, and result quantification, and evaluating them through experiments. Finally, this work presents the challenges of privacy auditing and its future direction.

       

    /

    返回文章
    返回