Loading Now

Summary of Smalltolarge (s2l): Scalable Data Selection For Fine-tuning Large Language Models by Summarizing Training Trajectories Of Small Models, By Yu Yang et al.


SmallToLarge (S2L): Scalable Data Selection for Fine-tuning Large Language Models by Summarizing Training Trajectories of Small Models

by Yu Yang, Siddhartha Mishra, Jeffrey N Chiang, Baharan Mirzasoleiman

First submitted to arxiv on: 12 Mar 2024

Categories

  • Main: Computation and Language (cs.CL)
  • Secondary: Artificial Intelligence (cs.AI); Machine Learning (cs.LG)

     Abstract of paper      PDF of paper


GrooveSquid.com Paper Summaries

GrooveSquid.com’s goal is to make artificial intelligence research accessible by summarizing AI papers in simpler terms. Each summary below covers the same AI paper, written at different levels of difficulty. The medium difficulty and low difficulty versions are original summaries written by GrooveSquid.com, while the high difficulty version is the paper’s original abstract. Feel free to learn from the version that suits you best!

Summary difficulty Written by Summary
High Paper authors High Difficulty Summary
Read the original abstract here
Medium GrooveSquid.com (original content) Medium Difficulty Summary
An effective and scalable data selection method for supervised fine-tuning (SFT) is introduced in this paper. The SmallToLarge (S2L) algorithm leverages training trajectories from small models to guide data selection for larger models, improving data efficiency in SFT for specialized domains such as mathematical problem-solving and clinical text summarization. Through extensive experiments, S2L significantly outperforms state-of-the-art data selection algorithms, reducing the training data required to match full dataset performance while achieving improved accuracy on challenging benchmarks like MATH and Phi-2. The algorithm’s ability to perform data selection using a reference model 40x smaller than the target model proportionally reduces the cost of data selection.
Low GrooveSquid.com (original content) Low Difficulty Summary
This paper introduces a new way to pick the right data for training language models. The method, called SmallToLarge (S2L), helps train larger models by learning from smaller models that have already been trained on less data. This approach makes it possible to achieve the same level of performance as training on all the available data, but using much less data. The results show that S2L outperforms other methods in various tasks, such as math problems and summarizing medical texts.

Keywords

* Artificial intelligence  * Fine tuning  * Summarization  * Supervised