Loading Now

Summary of Privacy-preserving Fine-tuning Of Large Language Models Through Flatness, by Tiejin Chen et al.


Privacy-preserving Fine-tuning of Large Language Models through Flatness

by Tiejin Chen, Longchao Da, Huixue Zhou, Pingzhi Li, Kaixiong Zhou, Tianlong Chen, Hua Wei

First submitted to arxiv on: 7 Mar 2024

Categories

  • Main: Artificial Intelligence (cs.AI)
  • Secondary: None

     Abstract of paper      PDF of paper


GrooveSquid.com Paper Summaries

GrooveSquid.com’s goal is to make artificial intelligence research accessible by summarizing AI papers in simpler terms. Each summary below covers the same AI paper, written at different levels of difficulty. The medium difficulty and low difficulty versions are original summaries written by GrooveSquid.com, while the high difficulty version is the paper’s original abstract. Feel free to learn from the version that suits you best!

Summary difficulty Written by Summary
High Paper authors High Difficulty Summary
Read the original abstract here
Medium GrooveSquid.com (original content) Medium Difficulty Summary
Recent advancements in Large Language Models (LLMs) have raised concerns about their privacy implications, particularly with the emergence of models like ChatGPT. To address these issues, researchers have explored Differential Privacy (DP) techniques to mitigate privacy risks at the cost of generalization degradation. Our paper reveals that the flatness of DP-trained models’ loss landscape plays a crucial role in the trade-off between privacy and generalization. We propose a holistic framework to enforce weight flatness, which improves model generalization while preserving competitive privacy. This framework innovates from three levels: perturbation-aware min-max optimization within layers, flatness-guided sparse prefix-tuning across layers, and weight knowledge distillation between DP and non-DP weights copies. Comprehensive experiments in both black-box and white-box scenarios demonstrate the effectiveness of our proposal in enhancing generalization and maintaining DP characteristics.
Low GrooveSquid.com (original content) Low Difficulty Summary
This paper is about making sure that language models are private and don’t leak personal information, while still being useful for tasks like text classification. It’s a problem because these models can learn to predict sensitive information about people just by looking at the words they use. The authors found that one way to make these models more private is to make the “loss landscape” of their training process flatter. This helps them be more accurate and generalizable, while still keeping personal info safe. They also proposed a new framework for doing this, which combines three different techniques to achieve better results.

Keywords

» Artificial intelligence  » Generalization  » Knowledge distillation  » Optimization  » Text classification