Summary of Few-shot Classification Of Interactive Activities Of Daily Living (interactadl), by Zane Durante et al.
Few-Shot Classification of Interactive Activities of Daily Living (InteractADL)
by Zane Durante, Robathan Harries, Edward Vendrow, Zelun Luo, Yuta Kyuragi, Kazuki Kozuka, Li Fei-Fei, Ehsan Adeli
First submitted to arxiv on: 3 Jun 2024
Categories
- Main: Computer Vision and Pattern Recognition (cs.CV)
- Secondary: Artificial Intelligence (cs.AI)
GrooveSquid.com Paper Summaries
GrooveSquid.com’s goal is to make artificial intelligence research accessible by summarizing AI papers in simpler terms. Each summary below covers the same AI paper, written at different levels of difficulty. The medium difficulty and low difficulty versions are original summaries written by GrooveSquid.com, while the high difficulty version is the paper’s original abstract. Feel free to learn from the version that suits you best!
Summary difficulty | Written by | Summary |
---|---|---|
High | Paper authors | High Difficulty Summary Read the original abstract here |
Medium | GrooveSquid.com (original content) | Medium Difficulty Summary A novel dataset and benchmark called InteractADL is proposed for understanding complex Activities of Daily Living (ADLs) that involve interaction between humans and objects. This paper addresses the challenges posed by these complex ADLs, including a long-tailed distribution due to the rarity of multi-person interactions and fine-grained visual recognition tasks due to semantically and visually similar classes. A novel method called Name Tuning is introduced for fine-grained few-shot video classification, which enables greater semantic separability by learning optimal class name vectors. This approach can be combined with existing prompt tuning strategies to learn the entire input text, demonstrating improved performance on InteractADL and other fine-grained visual classification benchmarks. |
Low | GrooveSquid.com (original content) | Low Difficulty Summary A new way of understanding how people do everyday things, like cooking or cleaning, is developed in this paper. The researchers created a special set of videos and rules (called a benchmark) to help machines learn these activities better. They also made a new method for teaching machines to recognize when someone is doing something specific, like washing dishes. This method works by learning the right words to describe what’s happening in the video. It can even work with just a few examples of each activity! The goal is to make machines more helpful and assistive in our daily lives. |
Keywords
» Artificial intelligence » Classification » Few shot » Prompt