Separation of Powers in Federated Learning

التفاصيل البيبلوغرافية
العنوان: Separation of Powers in Federated Learning
المؤلفون: Cheng, Pau-Chen, Eykholt, Kevin, Gu, Zhongshu, Jamjoom, Hani, Jayaram, K. R., Valdez, Enriquillo, Verma, Ashish
سنة النشر: 2021
المجموعة: Computer Science
مصطلحات موضوعية: Computer Science - Cryptography and Security, Computer Science - Machine Learning
الوصف: Federated Learning (FL) enables collaborative training among mutually distrusting parties. Model updates, rather than training data, are concentrated and fused in a central aggregation server. A key security challenge in FL is that an untrustworthy or compromised aggregation process might lead to unforeseeable information leakage. This challenge is especially acute due to recently demonstrated attacks that have reconstructed large fractions of training data from ostensibly "sanitized" model updates. In this paper, we introduce TRUDA, a new cross-silo FL system, employing a trustworthy and decentralized aggregation architecture to break down information concentration with regard to a single aggregator. Based on the unique computational properties of model-fusion algorithms, all exchanged model updates in TRUDA are disassembled at the parameter-granularity and re-stitched to random partitions designated for multiple TEE-protected aggregators. Thus, each aggregator only has a fragmentary and shuffled view of model updates and is oblivious to the model architecture. Our new security mechanisms can fundamentally mitigate training reconstruction attacks, while still preserving the final accuracy of trained models and keeping performance overheads low.
نوع الوثيقة: Working Paper
URL الوصول: http://arxiv.org/abs/2105.09400
رقم الأكسشن: edsarx.2105.09400
قاعدة البيانات: arXiv