TY - GEN
T1 - MAFD
T2 - 38th Annual ACM Symposium on Applied Computing, SAC 2023
AU - Wu, Aming
AU - Kwon, Young Woo
N1 - Publisher Copyright:
© 2023 Owner/Author(s).
PY - 2023/3/27
Y1 - 2023/3/27
N2 - The key challenges that recommendation systems must overcome are data isolation and privacy protection issues. Federated learning can efficiently train global models using decentralized data while preserving privacy. In real-world applications, however, it is difficult to achieve high prediction accuracy due to the heterogeneity of devices, the lack of data, and the limited generalization capacity of models. In this research, we introduce a personalized federated knowledge distillation model for a recommendation system based on a multi-head attention mechanism for recommendation systems. Specifically, we first employ federated distillation to improve the performance of student models and introduce a multi-head attention mechanism to enhance user encoding information. Next, we incorporate Wasserstein distance into the objective function of combined distillation to reduce the distribution gap between teacher and student networks and also use an adaptive learning rate technique to enhance convergence. We show that the proposed approach achieves better effectiveness and robustness through benchmarks.
AB - The key challenges that recommendation systems must overcome are data isolation and privacy protection issues. Federated learning can efficiently train global models using decentralized data while preserving privacy. In real-world applications, however, it is difficult to achieve high prediction accuracy due to the heterogeneity of devices, the lack of data, and the limited generalization capacity of models. In this research, we introduce a personalized federated knowledge distillation model for a recommendation system based on a multi-head attention mechanism for recommendation systems. Specifically, we first employ federated distillation to improve the performance of student models and introduce a multi-head attention mechanism to enhance user encoding information. Next, we incorporate Wasserstein distance into the objective function of combined distillation to reduce the distribution gap between teacher and student networks and also use an adaptive learning rate technique to enhance convergence. We show that the proposed approach achieves better effectiveness and robustness through benchmarks.
KW - federated learning
KW - multi-head attention
KW - recommendation systems
KW - wasserstein distance
UR - http://www.scopus.com/inward/record.url?scp=85162877031&partnerID=8YFLogxK
U2 - 10.1145/3555776.3577849
DO - 10.1145/3555776.3577849
M3 - Conference contribution
AN - SCOPUS:85162877031
T3 - Proceedings of the ACM Symposium on Applied Computing
SP - 1221
EP - 1224
BT - Proceedings of the 38th ACM/SIGAPP Symposium on Applied Computing, SAC 2023
PB - Association for Computing Machinery
Y2 - 27 March 2023 through 31 March 2023
ER -