Privacy Preserving Machine Learning in Healthcare

Authors

  • Venkata Raju, Sirisha Balla, Prasad Rayi

Keywords:

necessitates, nevertheless, optimization, safeguard, comprehensible

Abstract

The population of older individuals requiring care in Danish society is increasing.
This trend necessitates innovative and more efficient methods of functioning within the healthcare industry. Fall prevention is an issue requiring attention. To avoid falls, it is essential to evaluate the risk of falling among senior individuals. Should the danger be elevated, the implementation of additional assistance equipment or training programs may commence. Currently, the fall risk assessment is a manual and ineffective process. The optimization of this approach via the use of machine learning techniques delineates the scope of this thesis.

Machine learning employs extensive datasets throughout the training process. Under typical circumstances, this is not an issue; nevertheless, within the context of this thesis, the training data is confidential and so has sensitive characteristics. This complicates the optimization of machine learning and is the fundamental issue addressed in this thesis. What methods may be used to train machine learning algorithms on sensitive datasets?

This thesis assesses the use of federated learning for training machine learning algorithms on decentralized datasets. Additionally, it examines the use of encryption methods and differential privacy to safeguard machine learning models against the disclosure of sensitive information. Finally, it examines how intricate machine learning models, particularly deep learning models, might be rendered explainable to facilitate clearer communication of the findings to senior citizens.   
Numerous tests have shown that federated learning provides a technique for decentralized learning, although at the expense of model performance. Both encryption approaches and differential privacy may enhance the security of machine learning models against data leakage, but at the expense of model performance and complexity. Finally, it is shown how a technique known as SHAP may assist in calculating SHAP feature values, hence making machine learning models more comprehensible.

Downloads

Download data is not yet available.

References

Accuracy of noisy counting. https: //georgianpartners.shinyapps.io/interactive_counting/. Accessed: 2020-10-16. [2]Federated learning: Collaborative machine learning without centralized training data. https://ai.googleblog.com/2017/04/ federated-learning-collaborative.html. Accessed: 2020-10-16.

Hj_lpemiddelbasen. https://hmi-basen.dk/. Accessed: 2020-10-16.

Hvad er tv_rspor?? https://www.tvaerspor.dk/hvad-er-tvarspor/. Accessed: 2020-10-16.

Interpreting complex models with shap values. https://medium.com/@gabrieltseng/ interpreting-complex-models-with-shap-values-1c187db6ec83. Accessed: 2020-10-16.

Local vs. global di_erential privacy. https://desfontain.es/privacy/ local-global-differential-privacy.html. Accessed: 2020-10-16.

Udacity course on secure and private ai. https: //www.udacity.com/course/secure-and-private-ai--ud185? irclickid=VzXTSiQ0NxyORSgwUx0Mo3ERUkiQpKVtnzkY080&irgwc=1& utm_source=affiliate&utm_medium=ads_n&aff=259799. Accessed: 2020-10-16.

What is secure multi-party computation? https://medium.com/pytorch/ what-is-secure-multi-party-computation-8c875fb36ca5. Accessed: 2020-10-16.

Why one-hot encode data in machine learning? https://machinelearningmastery.com/ why-one-hot-encode-data-in-machine-learning/. Accessed: 2020-10-16.

Aarhus University. AIR (AI Rehabilitation). https://projekter.au.dk/air/, 2020. Online; accessed 24 September 2020.

Martin Abadi, Andy Chu, Ian Goodfellow, H Brendan McMahan, Ilya Mironov, Kunal Talwar, and Li Zhang. Deep learning with di_erential privacy. In Proceedings of the 2016 ACM SIGSAC Conference on Computer and Communications Security, pages 308{318, 2016.

Mohammad Al-Rubaie and J Morris Chang. Privacy-preserving machine learning: Threats and solutions. IEEE Security & Privacy, 17(2):49{58, 2019.

Eugene Bagdasaryan, Omid Poursaeed, and Vitaly Shmatikov. Di_erential privacy has disparate impact on model accuracy. In Advances in Neural Information Processing Systems, pages 15479{15488, 2019.

Aur_elien Bellet, Amaury Habrard, and Marc Sebban. A survey on metric learning for feature vectors and structured data. arXiv preprint arXiv:1306.6709, 2013.

R. Bhardwaj, A. R. Nambiar, and D. Dutta. A study of machine learning in healthcare. In 2017 IEEE 41st Annual Computer Software and Applications Conference (COMPSAC), volume 2, pages 236{241, 2017.

Peter Bjerregaard and K Juel. Middellevetid og d_delighed i danmark. Ugeskrift for Laeger, 155(50):4097{100, 1993.

A. Callahan and N. Shah. Machine learning in healthcare. 2017.

Cynthia Dwork, Aaron Roth, et al. The algorithmic foundations of di_erential privacy. Foundations and Trends in Theoretical Computer Science, 9(3-4):211{407, 2014.

Jianjiang Feng and Anil K Jain. Fingerprint reconstruction: from minutiae to phase. IEEE transactions on pattern analysis and machine intelligence, 33(2):209{223, 2010.

J Franken_eld. Arti_cial intelligence (ai). Investopedia udgivet, 13(6):19, 2019.

Downloads

Published

30.10.2024

How to Cite

Venkata Raju. (2024). Privacy Preserving Machine Learning in Healthcare. International Journal of Intelligent Systems and Applications in Engineering, 12(4), 5642 –. Retrieved from https://ijisae.org/index.php/IJISAE/article/view/7504

Issue

Section

Research Article