Loading Now

Summary of Flea: Addressing Data Scarcity and Label Skew in Federated Learning Via Privacy-preserving Feature Augmentation, by Tong Xia and Abhirup Ghosh and Xinchi Qiu and Cecilia Mascolo


FLea: Addressing Data Scarcity and Label Skew in Federated Learning via Privacy-preserving Feature Augmentation

by Tong Xia, Abhirup Ghosh, Xinchi Qiu, Cecilia Mascolo

First submitted to arxiv on: 4 Dec 2023

Categories

  • Main: Machine Learning (cs.LG)
  • Secondary: Cryptography and Security (cs.CR); Distributed, Parallel, and Cluster Computing (cs.DC)

     Abstract of paper      PDF of paper


GrooveSquid.com Paper Summaries

GrooveSquid.com’s goal is to make artificial intelligence research accessible by summarizing AI papers in simpler terms. Each summary below covers the same AI paper, written at different levels of difficulty. The medium difficulty and low difficulty versions are original summaries written by GrooveSquid.com, while the high difficulty version is the paper’s original abstract. Feel free to learn from the version that suits you best!

Summary difficulty Written by Summary
High Paper authors High Difficulty Summary
Read the original abstract here
Medium GrooveSquid.com (original content) Medium Difficulty Summary
This paper proposes a novel Federated Learning (FL) framework called FLea, addressing challenges in existing FL methods when dealing with scarce and label-skewed data. FLea incorporates a global feature buffer to mitigate local model drift, a feature augmentation approach based on mix-ups for local training, and an obfuscation method to enhance privacy. Experimental results demonstrate FLea’s superiority over state-of-the-art FL counterparts in 13 out of 18 settings, with improvements of over 5%, while concurrently mitigating privacy vulnerabilities.
Low GrooveSquid.com (original content) Low Difficulty Summary
FLea is a new way for devices to work together and learn without sharing their data. This helps solve problems when some devices have very little or imbalanced data, which makes it hard to train good models. FLea uses three main ideas: storing shared features from multiple devices, mixing up local training data to reduce overfitting, and hiding the source of shared features to keep them private. The results show that FLea works better than other FL methods in many cases.

Keywords

* Artificial intelligence  * Federated learning  * Overfitting