Privacy Protection Based on Federated Learning

Privacy Protection Based on Federated Learning

Authors

  • Bin Liu School of Information Technology, Mapua University, Manila, Philippines & School of Software Engineering, Xiamen University of Technology, Xiamen, China https://orcid.org/0009-0002-0490-2151
  • Eric B. Blancaflor School of Information Technology, Mapua University, Manila, Philippines https://orcid.org/0009-0002-0490-2151
  • Tianke Fang School of Computing and Information Engineering, Xiamen University of Technology, Xiamen, China https://orcid.org/0009-0009-2518-2502
  • Liming Cao School of Software Engineering, Xiamen University of Technology, Xiamen, China

DOI:

https://doi.org/10.37965/jait.2024.0503

Keywords:

federated learning, federal average algorithm, convolutional neural networks, privacy protection, privacy budget

Abstract

With the development of artificial intelligence technology, more and more fields will collect relevant user data, and provide users with a better experience through data analysis. But there are also risks involved in the process of data collection, namely how to protect personal privacy data. To address this issue, this study combined differential privacy, convolutional neural networks, and federated averaging algorithms to construct a privacy protection model. The study first utilized the federated average algorithm to handle data imbalance, ensuring that each analyzed data is in a balanced state. Then, based on of data balancing, a new algorithm model was constructed using differential privacy and convolutional neural networks. Finally, it utilized a number of public datasets to verify the role of the model in privacy protection. The results showed that the model can achieve recognition accuracy of 97.27% and 93.15%, respectively , for data under the influence of privacy budget and relaxation factor. Meanwhile, the classification accuracy of the model for data can reached 95.31%, with a regression error of 9.03%. When the local iteration number of the device was 30, the testing accuracy can reached 95.28%. This indicates that methods on the grounds of federated averaging algorithm and differential privacy can maintain the accuracy of the model while protecting user privacy. The application research of models has strong practical significance.

Metrics

Metrics Loading ...

Downloads

Published

2024-03-28

How to Cite

Liu, B., Eric B. Blancaflor, Fang, T., & Cao, L. (2024). Privacy Protection Based on Federated Learning. Journal of Artificial Intelligence and Technology. https://doi.org/10.37965/jait.2024.0503

Issue

Section

Research Articles
Loading...