Generative Models with Privacy Guarantees: Enhancing Data Utility while Minimizing Risk of Sensitive Data Exposure

Authors

  • Mohanarajesh Kommineni

Keywords:

effectiveness, integration, federated, computations

Abstract

The rapid advancement in generative models, including Generative Adversarial Networks (GANs), Variational Autoencoders (VAEs), and diffusion models, has significantly enhanced our ability to create high-quality synthetic data. These models have been instrumental in various applications, ranging from data augmentation and simulation to the development of privacy-preserving solutions. However, the generation of synthetic data also raises critical privacy concerns, as there is potential for these models to inadvertently reveal sensitive information about individuals in the original datasets. This paper delves into the intersection of generative models and data privacy, focusing on the development of techniques that safeguard privacy while ensuring the synthetic data produced remains meaningful and useful. We provide a comprehensive review of privacy-preserving strategies employed in the context of generative models. Key approaches discussed include differential privacy, which guarantees that the inclusion or exclusion of any individual data point does not significantly alter the output of a function; federated learning, which enables collaborative model training across decentralized data sources without sharing raw data; and secure multi-party computation (MPC), which allows for computations on encrypted data while preserving privacy. The paper evaluates these techniques in terms of their effectiveness, trade-offs, and integration challenges.

Downloads

Download data is not yet available.

References

Goodfellow et al., "Generative Adversarial Nets," Proceedings of the 27th International Conference on Neural Information Processing Systems (NIPS 2014), pp. 2672-2680, 2014.

D. P. Kingma and M. Welling, "Auto-Encoding Variational Bayes," Proceedings of the 2nd International Conference on Learning Representations (ICLR 2014), 2014.

J. Ho et al., "Denoising Diffusion Probabilistic Models," Proceedings of the 34th Conference on Neural Information Processing Systems (NeurIPS 2020), 2020.

S. Shokri and V. Shmatikov, "Privacy-Preserving Deep Learning," Proceedings of the 2015 ACM SIGSAC Conference on Computer and Communications Security (CCS 2015), pp. 1310-1321, 2015.

L. Melis et al., "Exploring the Limits of Differential Privacy for Federated Learning," Proceedings of the 2020 Conference on Neural Information Processing Systems (NeurIPS 2020), pp. 166-176, 2020.

C. Dwork et al., "Our Data, Ourselves: Privacy Via Distributed Noise Generation," Proceedings of the 2006 ACM SIGSAC Conference on Computer and Communications Security (CCS 2006), pp. 111-120, 2006.

J. Konečný, H. B. McMahan, and D. Ramage, "Federated Learning: Strategies for Improving Communication Efficiency," Proceedings of the 20th International Conference on Artificial Intelligence and Statistics (AISTATS 2017), vol. 54, pp. 268-277, 2017.

Y. Lindell and B. Pinkas, "Secure Multiparty Computation for Privacy-Preserving Data Analysis," IEEE Transactions on Computers, vol. 58, no. 5, pp. 799-812, 2009.

A. Abadi et al., "Deep Learning with Differential Privacy," Proceedings of the 2016 ACM SIGSAC Conference on Computer and Communications Security (CCS 2016), pp. 308-318, 2016.

M. H. M. and K. R., "Federated Learning for Healthcare Data: A Review," IEEE Reviews in Biomedical Engineering, vol. 14, pp. 151-164, 2021.

V. G. B. and G. T., "Secure Multi-Party Computation with Application to Machine Learning," IEEE Transactions on Information Forensics and Security, vol. 15, pp. 2389-2401, 2020.

Downloads

Published

06.08.2024

How to Cite

Mohanarajesh Kommineni. (2024). Generative Models with Privacy Guarantees: Enhancing Data Utility while Minimizing Risk of Sensitive Data Exposure. International Journal of Intelligent Systems and Applications in Engineering, 12(23s), 1036–1044. Retrieved from https://ijisae.org/index.php/IJISAE/article/view/7111

Issue

Section

Research Article