Loading Now

Summary of Harvesting Private Medical Images in Federated Learning Systems with Crafted Models, by Shanghao Shi et al.


Harvesting Private Medical Images in Federated Learning Systems with Crafted Models

by Shanghao Shi, Md Shahedul Haque, Abhijeet Parida, Marius George Linguraru, Y.Thomas Hou, Syed Muhammad Anwar, Wenjing Lou

First submitted to arxiv on: 13 Jul 2024

Categories

  • Main: Machine Learning (cs.LG)
  • Secondary: Cryptography and Security (cs.CR); Image and Video Processing (eess.IV)

     Abstract of paper      PDF of paper


GrooveSquid.com Paper Summaries

GrooveSquid.com’s goal is to make artificial intelligence research accessible by summarizing AI papers in simpler terms. Each summary below covers the same AI paper, written at different levels of difficulty. The medium difficulty and low difficulty versions are original summaries written by GrooveSquid.com, while the high difficulty version is the paper’s original abstract. Feel free to learn from the version that suits you best!

Summary difficulty Written by Summary
High Paper authors High Difficulty Summary
Read the original abstract here
Medium GrooveSquid.com (original content) Medium Difficulty Summary
The proposed novel attack, MediLeak, targets federated learning (FL) by allowing a malicious parameter server to recover high-fidelity patient images from model updates uploaded by clients. This is achieved by generating an adversarial model with a crafted module and having each client train on it, sending the resulting model updates back to the server. The attack exploits FL protocols and can break state-of-the-art cryptographic secure aggregation protocols designed to protect FL systems from privacy inference attacks. MediLeak was tested on MedMNIST and COVIDx CXR-4 datasets, achieving nearly perfect recovery of private images with high recovery rates and quantitative scores. Furthermore, downstream tasks such as disease classification showed no significant performance degradation compared to using original training samples.
Low GrooveSquid.com (original content) Low Difficulty Summary
A new kind of attack called MediLeak threatens the security of a type of machine learning called federated learning (FL). FL lets many devices work together without sharing their own data. But MediLeak can make these devices share their data anyway by tricking them into training on a special model. This attack is bad because it means people’s private information, like medical images, could be leaked. The researchers tested MediLeak and found that it was very good at getting back the original images from the device updates. They also did some extra tasks with the recovered data and saw no significant difference in how well they worked.

Keywords

» Artificial intelligence  » Classification  » Federated learning  » Inference  » Machine learning