Loading Now

Summary of The Problem Of Social Cost in Multi-agent General Reinforcement Learning: Survey and Synthesis, by Kee Siong Ng et al.


The Problem of Social Cost in Multi-Agent General Reinforcement Learning: Survey and Synthesis

by Kee Siong Ng, Samuel Yang-Zhao, Timothy Cadogan-Cowper

First submitted to arxiv on: 3 Dec 2024

Categories

  • Main: Artificial Intelligence (cs.AI)
  • Secondary: Computer Science and Game Theory (cs.GT); Machine Learning (cs.LG); Multiagent Systems (cs.MA)

     Abstract of paper      PDF of paper


GrooveSquid.com Paper Summaries

GrooveSquid.com’s goal is to make artificial intelligence research accessible by summarizing AI papers in simpler terms. Each summary below covers the same AI paper, written at different levels of difficulty. The medium difficulty and low difficulty versions are original summaries written by GrooveSquid.com, while the high difficulty version is the paper’s original abstract. Feel free to learn from the version that suits you best!

Summary difficulty Written by Summary
High Paper authors High Difficulty Summary
Read the original abstract here
Medium GrooveSquid.com (original content) Medium Difficulty Summary
The paper proposes market-based mechanisms to quantify and control social harms in multi-agent environments where artificial general intelligence (AGI) agents interact. This is a crucial issue in AI safety, as AGI agents may unintentionally cause catastrophic collateral damage while optimizing specific objectives. The proposed setup builds upon existing formulations of multi-agent reinforcement learning with mechanism design, but adds two key features: a history-based general reinforcement learning environment like AIXI and the ability for agents to have different learning strategies and planning horizons. To demonstrate the practicality of this approach, the paper surveys various learning algorithms and presents applications such as the Paperclips problem and pollution control using cap-and-trade systems.
Low GrooveSquid.com (original content) Low Difficulty Summary
The paper solves a big problem in AI safety by finding a way to measure and control the bad things that can happen when artificial intelligence agents work together. Right now, these agents might accidentally cause a lot of harm while trying to achieve their goals. The solution involves creating a system where agents are rewarded or punished based on how well they behave. This approach is more realistic than current methods because it allows agents to have different ways of learning and planning for the future. The paper shows that this method can be used in real-world scenarios, such as managing pollution.

Keywords

» Artificial intelligence  » Reinforcement learning