Generative Models with Privacy Guarantees: Enhancing Data Utility while Minimizing Risk of Sensitive Data Exposure
Keywords:
effectiveness, integration, federated, computationsAbstract
The rapid advancement in generative models, including Generative Adversarial Networks (GANs), Variational Autoencoders (VAEs), and diffusion models, has significantly enhanced our ability to create high-quality synthetic data. These models have been instrumental in various applications, ranging from data augmentation and simulation to the development of privacy-preserving solutions. However, the generation of synthetic data also raises critical privacy concerns, as there is potential for these models to inadvertently reveal sensitive information about individuals in the original datasets. This paper delves into the intersection of generative models and data privacy, focusing on the development of techniques that safeguard privacy while ensuring the synthetic data produced remains meaningful and useful. We provide a comprehensive review of privacy-preserving strategies employed in the context of generative models. Key approaches discussed include differential privacy, which guarantees that the inclusion or exclusion of any individual data point does not significantly alter the output of a function; federated learning, which enables collaborative model training across decentralized data sources without sharing raw data; and secure multi-party computation (MPC), which allows for computations on encrypted data while preserving privacy. The paper evaluates these techniques in terms of their effectiveness, trade-offs, and integration challenges.
Downloads
References
Goodfellow et al., "Generative Adversarial Nets," Proceedings of the 27th International Conference on Neural Information Processing Systems (NIPS 2014), pp. 2672-2680, 2014.
D. P. Kingma and M. Welling, "Auto-Encoding Variational Bayes," Proceedings of the 2nd International Conference on Learning Representations (ICLR 2014), 2014.
J. Ho et al., "Denoising Diffusion Probabilistic Models," Proceedings of the 34th Conference on Neural Information Processing Systems (NeurIPS 2020), 2020.
S. Shokri and V. Shmatikov, "Privacy-Preserving Deep Learning," Proceedings of the 2015 ACM SIGSAC Conference on Computer and Communications Security (CCS 2015), pp. 1310-1321, 2015.
L. Melis et al., "Exploring the Limits of Differential Privacy for Federated Learning," Proceedings of the 2020 Conference on Neural Information Processing Systems (NeurIPS 2020), pp. 166-176, 2020.
C. Dwork et al., "Our Data, Ourselves: Privacy Via Distributed Noise Generation," Proceedings of the 2006 ACM SIGSAC Conference on Computer and Communications Security (CCS 2006), pp. 111-120, 2006.
J. Konečný, H. B. McMahan, and D. Ramage, "Federated Learning: Strategies for Improving Communication Efficiency," Proceedings of the 20th International Conference on Artificial Intelligence and Statistics (AISTATS 2017), vol. 54, pp. 268-277, 2017.
Y. Lindell and B. Pinkas, "Secure Multiparty Computation for Privacy-Preserving Data Analysis," IEEE Transactions on Computers, vol. 58, no. 5, pp. 799-812, 2009.
A. Abadi et al., "Deep Learning with Differential Privacy," Proceedings of the 2016 ACM SIGSAC Conference on Computer and Communications Security (CCS 2016), pp. 308-318, 2016.
M. H. M. and K. R., "Federated Learning for Healthcare Data: A Review," IEEE Reviews in Biomedical Engineering, vol. 14, pp. 151-164, 2021.
V. G. B. and G. T., "Secure Multi-Party Computation with Application to Machine Learning," IEEE Transactions on Information Forensics and Security, vol. 15, pp. 2389-2401, 2020.
Downloads
Published
How to Cite
Issue
Section
License
This work is licensed under a Creative Commons Attribution-ShareAlike 4.0 International License.
All papers should be submitted electronically. All submitted manuscripts must be original work that is not under submission at another journal or under consideration for publication in another form, such as a monograph or chapter of a book. Authors of submitted papers are obligated not to submit their paper for publication elsewhere until an editorial decision is rendered on their submission. Further, authors of accepted papers are prohibited from publishing the results in other publications that appear before the paper is published in the Journal unless they receive approval for doing so from the Editor-In-Chief.
IJISAE open access articles are licensed under a Creative Commons Attribution-ShareAlike 4.0 International License. This license lets the audience to give appropriate credit, provide a link to the license, and indicate if changes were made and if they remix, transform, or build upon the material, they must distribute contributions under the same license as the original.