Loading Now

Summary of Semi-supervised One-shot Imitation Learning, by Philipp Wu and Kourosh Hakhamaneshi and Yuqing Du and Igor Mordatch and Aravind Rajeswaran and Pieter Abbeel


Semi-Supervised One-Shot Imitation Learning

by Philipp Wu, Kourosh Hakhamaneshi, Yuqing Du, Igor Mordatch, Aravind Rajeswaran, Pieter Abbeel

First submitted to arxiv on: 9 Aug 2024

Categories

  • Main: Machine Learning (cs.LG)
  • Secondary: Artificial Intelligence (cs.AI)

     Abstract of paper      PDF of paper


GrooveSquid.com Paper Summaries

GrooveSquid.com’s goal is to make artificial intelligence research accessible by summarizing AI papers in simpler terms. Each summary below covers the same AI paper, written at different levels of difficulty. The medium difficulty and low difficulty versions are original summaries written by GrooveSquid.com, while the high difficulty version is the paper’s original abstract. Feel free to learn from the version that suits you best!

Summary difficulty Written by Summary
High Paper authors High Difficulty Summary
Read the original abstract here
Medium GrooveSquid.com (original content) Medium Difficulty Summary
The paper introduces One-shot Imitation Learning (OSIL), which enables AI agents to learn from a single demonstration. Typically, OSIL requires numerous paired expert demonstrations, but this can be impractical. To overcome this limitation, the authors propose a semi-supervised OSIL setting, where the agent receives an unpaired dataset of trajectories and a small paired dataset with task labels. This setting mimics few-shot learning, requiring the agent to leverage weak supervision from the large dataset. The authors develop an algorithm for this setting, first learning an embedding space that clusters tasks uniquely. They then use this space to self-generate pairings between unpaired trajectories, allowing OSIL models to be trained with competitive performance using only weak supervision.
Low GrooveSquid.com (original content) Low Difficulty Summary
OSIL is a way to teach AI agents new skills by showing them one example. Usually, you need many examples to train an AI agent, but that’s not always possible. In this paper, the authors solve this problem by giving the agent a big collection of actions (trajectories) with no labels and a small group of labeled examples. This helps the agent learn more efficiently and accurately.

Keywords

» Artificial intelligence  » Embedding space  » Few shot  » One shot  » Semi supervised