Loading Now

Summary of Q-adapter: Customizing Pre-trained Llms to New Preferences with Forgetting Mitigation, by Yi-chen Li et al.


Q-Adapter: Customizing Pre-trained LLMs to New Preferences with Forgetting Mitigation

by Yi-Chen Li, Fuxiang Zhang, Wenjie Qiu, Lei Yuan, Chengxing Jia, Zongzhang Zhang, Yang Yu, Bo An

First submitted to arxiv on: 4 Jul 2024

Categories

  • Main: Machine Learning (cs.LG)
  • Secondary: None

     Abstract of paper      PDF of paper


GrooveSquid.com Paper Summaries

GrooveSquid.com’s goal is to make artificial intelligence research accessible by summarizing AI papers in simpler terms. Each summary below covers the same AI paper, written at different levels of difficulty. The medium difficulty and low difficulty versions are original summaries written by GrooveSquid.com, while the high difficulty version is the paper’s original abstract. Feel free to learn from the version that suits you best!

Summary difficulty Written by Summary
High Paper authors High Difficulty Summary
Read the original abstract here
Medium GrooveSquid.com (original content) Medium Difficulty Summary
The proposed Q-Adapter method allows for the customization of pre-trained Large Language Models (LLMs) to new human preferences while preserving their original capabilities. This is achieved by casting LLM customization as optimizing a sum of two reward functions: one used during pre-training and another characterizing the new preference. The approach leverages the residual Q-learning framework, enabling the restoration of customized LLMs without requiring knowledge of the reward function. Experiments demonstrate the effectiveness of Q-Adapter on retaining existing knowledge and learning new preferences using the Llama-3.1 model and datasets such as DSP and HH-RLHF.
Low GrooveSquid.com (original content) Low Difficulty Summary
The paper proposes a way to make Large Language Models (LLMs) do what humans want, while still keeping their original abilities. This is done by changing how the LLMs are trained to fit new preferences. The method uses a special kind of learning called residual Q-learning and a module called an adapter. It works well on real-world datasets like DSP and HH-RLHF.

Keywords

» Artificial intelligence  » Llama  » Rlhf