Loading Now

Summary of Large Language Monkeys: Scaling Inference Compute with Repeated Sampling, by Bradley Brown et al.


Large Language Monkeys: Scaling Inference Compute with Repeated Sampling

by Bradley Brown, Jordan Juravsky, Ryan Ehrlich, Ronald Clark, Quoc V. Le, Christopher Ré, Azalia Mirhoseini

First submitted to arxiv on: 31 Jul 2024

Categories

  • Main: Machine Learning (cs.LG)
  • Secondary: Artificial Intelligence (cs.AI)

     Abstract of paper      PDF of paper


GrooveSquid.com Paper Summaries

GrooveSquid.com’s goal is to make artificial intelligence research accessible by summarizing AI papers in simpler terms. Each summary below covers the same AI paper, written at different levels of difficulty. The medium difficulty and low difficulty versions are original summaries written by GrooveSquid.com, while the high difficulty version is the paper’s original abstract. Feel free to learn from the version that suits you best!

Summary difficulty Written by Summary
High Paper authors High Difficulty Summary
Read the original abstract here
Medium GrooveSquid.com (original content) Medium Difficulty Summary
This paper explores the concept of inference compute in language models. Traditionally, models are limited to making one attempt at solving a problem during inference. The authors propose repeatedly sampling candidate solutions from a model to increase coverage, which is defined as the fraction of problems that can be solved by any generated sample. Across multiple tasks and models, the authors observe that coverage scales with the number of samples over four orders of magnitude, following a log-linear relationship. This paper highlights the importance of inference compute in scaling language models’ capabilities. The authors demonstrate improved performance in domains like coding and formal proofs where answers can be automatically verified. They also investigate common methods for picking from a sample collection, such as majority voting and reward models, which plateau beyond several hundred samples.
Low GrooveSquid.com (original content) Low Difficulty Summary
This paper is about improving how language models work during the process of solving problems. Normally, these models only try once to solve a problem. The authors suggest trying multiple times to find solutions. They found that when they do this, it helps solve more problems across different tasks and models. This can be useful in areas where answers can be easily checked, like coding and math proofs. The paper also looks at how people usually choose the best solution from a set of options and finds that these methods stop working well after trying many times.

Keywords

* Artificial intelligence  * Inference