Squeezing More Utility via Adaptive Clipping on Differentially Private Gradients in Federated Meta-Learning. Wang, N., Xiao, Y., Chen, Y., Zhang, N., Lou, W., & Hou, Y. T. In Proceedings of the 38th Annual Computer Security Applications Conference, of ACSAC '22, pages 647–657, New York, NY, USA, December, 2022. Association for Computing Machinery.
Squeezing More Utility via Adaptive Clipping on Differentially Private Gradients in Federated Meta-Learning [link]Paper  doi  abstract   bibtex   
Federated meta-learning has emerged as a promising AI framework for today’s mobile computing scenes involving distributed clients. It enables collaborative model training using the data located at distributed mobile clients and accommodates clients that need fast model customization with limited new data. However, federated meta-learning solutions are susceptible to inference-based privacy attacks since the global model encoded with clients’ training data is open to all clients and the central server. Meanwhile, differential privacy (DP) has been widely used as a countermeasure against privacy inference attacks in federated learning. The adoption of DP in federated meta-learning is complicated by the model accuracy-privacy trade-off and the model hierarchy attributed to the meta-learning component. In this paper, we introduce DP-FedMeta, a new differentially private federated meta-learning architecture that addresses such data privacy challenges. DP-FedMeta features an adaptive gradient clipping method and a one-pass meta-training process to improve the model utility-privacy trade-off. At the core of DP-FedMeta are two DP mechanisms, namely DP-AGR and DP-AGRLR, to provide two notions of privacy protection for the hierarchical models. Extensive experiments in an emulated federated meta-learning scenario on well-known datasets (Omniglot, CIFAR-FS, and Mini-ImageNet) demonstrate that DP-FedMeta accomplishes better privacy protection while maintaining comparable model accuracy compared to the state-of-the-art solution that directly applies DP-based meta-learning to the federated setting.
@inproceedings{wang_squeezing_2022,
	address = {New York, NY, USA},
	series = {{ACSAC} '22},
	title = {Squeezing {More} {Utility} via {Adaptive} {Clipping} on {Differentially} {Private} {Gradients} in {Federated} {Meta}-{Learning}},
	isbn = {9781450397599},
	url = {https://dl.acm.org/doi/10.1145/3564625.3564652},
	doi = {10.1145/3564625.3564652},
	abstract = {Federated meta-learning has emerged as a promising AI framework for today’s mobile computing scenes involving distributed clients. It enables collaborative model training using the data located at distributed mobile clients and accommodates clients that need fast model customization with limited new data. However, federated meta-learning solutions are susceptible to inference-based privacy attacks since the global model encoded with clients’ training data is open to all clients and the central server. Meanwhile, differential privacy (DP) has been widely used as a countermeasure against privacy inference attacks in federated learning. The adoption of DP in federated meta-learning is complicated by the model accuracy-privacy trade-off and the model hierarchy attributed to the meta-learning component. In this paper, we introduce DP-FedMeta, a new differentially private federated meta-learning architecture that addresses such data privacy challenges. DP-FedMeta features an adaptive gradient clipping method and a one-pass meta-training process to improve the model utility-privacy trade-off. At the core of DP-FedMeta are two DP mechanisms, namely DP-AGR and DP-AGRLR, to provide two notions of privacy protection for the hierarchical models. Extensive experiments in an emulated federated meta-learning scenario on well-known datasets (Omniglot, CIFAR-FS, and Mini-ImageNet) demonstrate that DP-FedMeta accomplishes better privacy protection while maintaining comparable model accuracy compared to the state-of-the-art solution that directly applies DP-based meta-learning to the federated setting.},
	urldate = {2024-02-08},
	booktitle = {Proceedings of the 38th {Annual} {Computer} {Security} {Applications} {Conference}},
	publisher = {Association for Computing Machinery},
	author = {Wang, Ning and Xiao, Yang and Chen, Yimin and Zhang, Ning and Lou, Wenjing and Hou, Y. Thomas},
	month = dec,
	year = {2022},
	keywords = {adaptive clipping, differential privacy, federated meta-learning, privacy utility trade-off},
	pages = {647--657},
}

Downloads: 0