Loading Now

Summary of Mastering Board Games by External and Internal Planning with Language Models, By John Schultz et al.


Mastering Board Games by External and Internal Planning with Language Models

by John Schultz, Jakub Adamek, Matej Jusup, Marc Lanctot, Michael Kaisers, Sarah Perrin, Daniel Hennes, Jeremy Shar, Cannada Lewis, Anian Ruoss, Tom Zahavy, Petar Veličković, Laurel Prince, Satinder Singh, Eric Malmi, Nenad Tomašev

First submitted to arxiv on: 2 Dec 2024

Categories

  • Main: Artificial Intelligence (cs.AI)
  • Secondary: Computation and Language (cs.CL); Machine Learning (cs.LG)

     Abstract of paper      PDF of paper


GrooveSquid.com Paper Summaries

GrooveSquid.com’s goal is to make artificial intelligence research accessible by summarizing AI papers in simpler terms. Each summary below covers the same AI paper, written at different levels of difficulty. The medium difficulty and low difficulty versions are original summaries written by GrooveSquid.com, while the high difficulty version is the paper’s original abstract. Feel free to learn from the version that suits you best!

Summary difficulty Written by Summary
High Paper authors High Difficulty Summary
Read the original abstract here
Medium GrooveSquid.com (original content) Medium Difficulty Summary
A novel approach combines large language models (LLMs) with search-based planning to significantly improve their performance in complex tasks such as multi-step reasoning. The authors introduce two methods: external search, which uses Monte Carlo Tree Search (MCTS) guided by the LLM without an external engine, and internal search, which generates a linearized tree of potential futures and a final choice within the LLM. Both approaches build on a pre-trained language model with domain knowledge, capturing transition and value functions across various board games (Chess, Fischer Random/Chess960, Connect Four, and Hex). The pre-training method minimizes hallucinations by accurately predicting states and legal moves. Results show that both internal and external search improve win-rates against state-of-the-art bots, even reaching Grandmaster-level performance in chess while operating on a similar move count search budget per decision as human Grandmasters.
Low GrooveSquid.com (original content) Low Difficulty Summary
Large language models can play games like chess or Connect Four well, but they struggle with planning and reasoning. Scientists found a way to make these models better at planning by combining them with special kinds of searches. They tested two methods: one that uses an external search tool and another that does the searching inside the model itself. Both methods used language models trained on information about the games. The results showed that both approaches made the models much better at playing the games, even beating grandmasters in chess.

Keywords

» Artificial intelligence  » Language model