Abstract
Due to the difficulties of exchanging data securely, data silos have become a critical issue in the era of big data. Federated learning provides an advantageous approach by enabling data holders to train a model collaboratively without sharing local data. However, multiple known inference attacks have made it impossible for a purely federated learning approach to protect privacy well enough. We present a PBPAFL algorithm that combines differential privacy with homomorphic encryption based on federated learning with an assessment module that enables the privacy budget parameters to be flexible in response to varying training requirements. The models trained using our proposed PBPAFL algorithm are capable of preventing inference assaults without a severe loss of precision. To demonstrate the efficacy of our proposed framework, we employ the PBPAFL algorithm to train a collection of face image-sensitive data. The experimental results show that our approach can improve the privacy protection of the model while maintaining precision.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Colesky, M., Demetzou, K., Fritsch, L., Herold, S.: Helping software architects familiarize with the general data protection regulation. In: 2019 IEEE International Conference on Software Architecture Companion (ICSA-C), pp. 226–229. IEEE, United States (2019)
Kim, J., Ha, H., Chun, B.G., Yoon, S., Cha, S.K.: Collaborative analytics for data silos. In: 2016 IEEE 32nd International Conference on Data Engineering (ICDE), pp. 743–754. IEEE, United States (2016)
Ahmed, K.M., Imteaj, A., Amini, M.H.: Federated deep learning for heterogeneous edge computing. In: 2021 20th IEEE International Conference on Machine Learning and Applications (ICMLA), pp. 1146–1152. IEEE, United States (2021)
Song, C., Ristenpart, T., Shmatikov, V.: Machine learning models that remember too much. In: Proceedings of the 2017 ACM SIGSAC Conference on computer and communications security, pp. 587–601. ACM, United States (2017)
Zhu, L., Liu, Z., Han, S.: Deep leakage from gradients. Ion: Advances in Neural Information Processing Systems, vol. 32. NeurIPS, Vancouver (2019)
Dwork, C.: Differential privacy: a survey of results. In: Agrawal, M., Du, D., Duan, Z., Li, A. (eds.) TAMC 2008. LNCS, vol. 4978, pp. 1–19. Springer, Heidelberg (2008). https://doi.org/10.1007/978-3-540-79228-4_1
Bayardo, R.J., Agrawal, R.: Data privacy through optimal k-anonymization. In: 21st International Conference on Data Engineering (ICDE 2005), pp. 217–228. IEEE, United States (2005)
Gentry, C.: Computing arbitrary functions of encrypted data. Commun. ACM 53, 97–105 (2010)
Gentry, C., Halevi, S.: Implementing gentry’s fully-homomorphic encryption scheme. In: Paterson, K.G. (ed.) Advances in Cryptology – EUROCRYPT 2011. EUROCRYPT 2011. Lecture Notes in Computer Science, vol. 6632. Springer, Heidelberg (2011). https://doi.org/10.1007/978-3-642-20465-4_9
Acar, A., Aksu, H., Uluagac, A.S., Conti, M.: A survey on homomorphic encryption schemes: theory and implementation. ACM Comput. Surv. (CSUR) 51(4), 1–35 (2018)
Wang, F., Casalino, L.P., Khullar, D.: Deep learning in medicine—promise, progress, and challenges. JAMA Intern. Med. 179(3), 293–294 (2019)
Yaji, S., Bangera, K., Neelima, B.: Privacy preserving in blockchain based on partial homomorphic encryption system for AI applications. In: 2018 IEEE 25th International Conference on High Performance Computing Workshops (HiPCW), pp. 81–85. IEEE, United States (2018)
Li, X., Gu, Y., Dvornek, N., Staib, L.H., Ventola, P., Duncan, J.S.: Multi-site fMRI analysis using privacy-preserving federated learning and domain adaptation: ABIDE results. Med. Image Anal. 65, 101765 (2020)
Acknowledgment
We appreciate the informative remarks made by the anonymous reviewers of this work. This research is supported by the Special Fund for the Key Program of Science and Technology of Guangdong Province, China (Grant No. 2016B030305003).
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 ICST Institute for Computer Sciences, Social Informatics and Telecommunications Engineering
About this paper
Cite this paper
Yao, R., Tang, K., Zhu, Y., Fan, B., Luo, T., Song, Y. (2023). PBPAFL: A Federated Learning Framework with Hybrid Privacy Protection for Sensitive Data. In: Goel, S., Gladyshev, P., Nikolay, A., Markowsky, G., Johnson, D. (eds) Digital Forensics and Cyber Crime. ICDF2C 2022. Lecture Notes of the Institute for Computer Sciences, Social Informatics and Telecommunications Engineering, vol 508. Springer, Cham. https://doi.org/10.1007/978-3-031-36574-4_24
Download citation
DOI: https://doi.org/10.1007/978-3-031-36574-4_24
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-36573-7
Online ISBN: 978-3-031-36574-4
eBook Packages: Computer ScienceComputer Science (R0)