Federated Learning Shields Personal Data in AI Advancements

Abstract

Albeit the success of federated learning (FL) in decentralized training, bolstering the generalization of models by overcoming heterogeneity across clients still remains a huge challenge. To aim at improved generalization of FL, a group of recent works pursues flatter minima of models by employing sharpness-aware minimization in the local training at the client side. However, we observe that the global model, i.e., the aggregated model, does not lie on flat minima of the global objective, even with the effort of flatness searching in local training, which we define as flatness discrepancy. By rethinking and theoretically analyzing flatness searching in FL through the lens of the discrepancy problem, we propose a method called Federated Learning for Global Flatness (FedGF) that explicitly pursues the flatter minima of the global models, leading to the relieved flatness discrepancy and remarkable performance gains in the heterogeneous FL benchmarks.

With the advancement of artificial intelligence (AI), the issue of user privacy infringement has become increasingly serious. Federated learning (FL) has emerged as a core technology to address this challenge, especially in the realm of on-device AI that has garnered significant interest from IT companies.

Professor Sung Whan Yoon and his team in the Graduate School of Artificial Intelligence at UNIST have developed a groundbreaking technology, called FedGF (Federated Learning for Global Flatness) that enhances AI performance while safeguarding personal information. This innovation is poised to tackle the problem of user data leakage effectively.

The research team has created a method that consistently delivers high performance across diverse user data distributions. Unlike existing technologies, which perform well only in environments similar to the training data, FedGF maintains its efficacy across various scenarios.

Federated learning (FL) protects personal information by training deep learning models directly on user devices. However, its effectiveness can be hampered by data discrepancies among devices. FedGF achieves high accuracy by optimizing models based on the local training from each device, all without transmitting sensitive data to a central server.

In addition to its robustness, FedGF is highly efficient, requiring fewer communication resources than traditional methods. This efficiency is particularly beneficial for mobile devices that rely on wireless communication, such as Wi-Fi.

Professor Yoon stated, "Federated learning will serve as a crucial foundation in addressing AI privacy infringement." He added, "It will significantly assist major IT companies in overcoming challenges related to personal information and data heterogeneity."

First author Taehwan Lee remarked, "With FedGF technology, companies can develop high-performance AI models without infringing on personal information, thus paving the way for advancements in various fields such as IT, healthcare, and autonomous driving."

The research findings were published online on July 20 at the International Conference on Machine Learning (ICML). This project was conducted as part of an international collaboration on information protection, supported by the Ministry of Science and ICT (MSIT), as well as an initiative for training talent in the fields of information and communication broadcasting.

Journal Reference

Taehwan Lee and Sung Whan Yoon, "Rethinking the Flat Minima Searching in Federated Learning," ICML, (2024).

/Public Release. This material from the originating organization/author(s) might be of the point-in-time nature, and edited for clarity, style and length. Mirage.News does not take institutional positions or sides, and all views, positions, and conclusions expressed herein are solely those of the author(s).View in full here.