Loading Now

Summary of Towards Hyper-parameter-free Federated Learning, by Geetika et al.


Towards Hyper-parameter-free Federated Learning

by Geetika, Drishya Uniyal, Bapi Chatterjee

First submitted to arxiv on: 30 Aug 2024

Categories

  • Main: Machine Learning (cs.LG)
  • Secondary: Artificial Intelligence (cs.AI); Optimization and Control (math.OC)

     Abstract of paper      PDF of paper


GrooveSquid.com Paper Summaries

GrooveSquid.com’s goal is to make artificial intelligence research accessible by summarizing AI papers in simpler terms. Each summary below covers the same AI paper, written at different levels of difficulty. The medium difficulty and low difficulty versions are original summaries written by GrooveSquid.com, while the high difficulty version is the paper’s original abstract. Feel free to learn from the version that suits you best!

Summary difficulty Written by Summary
High Paper authors High Difficulty Summary
Read the original abstract here
Medium GrooveSquid.com (original content) Medium Difficulty Summary
The proposed automated scaling techniques in federated learning (FL) for global model updates outperform vanilla federated averaging (FedAvg), but existing methods require additional tunable hyperparameters on the server. The authors introduce two algorithms that eliminate the need for these hyperparameters, achieving competitive convergence rates and good empirical performance. The first algorithm uses a descent-ensuring step-size regime at clients to ensure linear convergence for strongly convex federated objectives. The second algorithm replaces the objective function value at the server with the average of objective values from sampled clients, facilitating computation of the scaling factor. Empirical results demonstrate that these methods perform similarly or better than popular FL algorithms for both convex and non-convex problems.
Low GrooveSquid.com (original content) Low Difficulty Summary
Federated learning is a way to train models on many devices without sharing all their data. The problem is that the updates from each device can be very different, making it hard to combine them correctly. Some methods try to fix this by adjusting the amount of update sent from each device, but these adjustments require careful tuning. This paper proposes two new algorithms that automatically adjust these updates, allowing for better performance without needing manual tweaking. The results show that these algorithms work well for a variety of problems and can be used with different types of models.

Keywords

» Artificial intelligence  » Federated learning  » Objective function